Dec 10 18:55:16 crc systemd[1]: Starting Kubernetes Kubelet... Dec 10 18:55:16 crc restorecon[4688]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:16 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:17 crc restorecon[4688]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 10 18:55:17 crc kubenswrapper[4894]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:17 crc kubenswrapper[4894]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 10 18:55:17 crc kubenswrapper[4894]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:17 crc kubenswrapper[4894]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:17 crc kubenswrapper[4894]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 10 18:55:17 crc kubenswrapper[4894]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.332828 4894 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340680 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340726 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340736 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340745 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340753 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340762 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340771 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340779 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340788 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340795 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340803 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340811 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340818 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340826 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340834 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340841 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340882 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340891 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340898 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340906 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340914 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340921 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340929 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340936 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340944 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340951 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340959 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340966 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340974 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340982 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340990 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.340998 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341006 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341014 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341022 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341031 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341039 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341046 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341053 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341063 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341074 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341083 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341091 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341099 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341107 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341116 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341123 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341131 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341143 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341153 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341163 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341171 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341180 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341189 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341197 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341206 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341214 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341222 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341231 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341242 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341253 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341261 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341270 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341278 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341286 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341294 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341302 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341310 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341318 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341330 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.341338 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341483 4894 flags.go:64] FLAG: --address="0.0.0.0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341502 4894 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341524 4894 flags.go:64] FLAG: --anonymous-auth="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341536 4894 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341548 4894 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341558 4894 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341570 4894 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341581 4894 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341594 4894 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341606 4894 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341619 4894 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341637 4894 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341649 4894 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341659 4894 flags.go:64] FLAG: --cgroup-root="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341668 4894 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341678 4894 flags.go:64] FLAG: --client-ca-file="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341687 4894 flags.go:64] FLAG: --cloud-config="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341696 4894 flags.go:64] FLAG: --cloud-provider="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341705 4894 flags.go:64] FLAG: --cluster-dns="[]" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341717 4894 flags.go:64] FLAG: --cluster-domain="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341727 4894 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341737 4894 flags.go:64] FLAG: --config-dir="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341746 4894 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341755 4894 flags.go:64] FLAG: --container-log-max-files="5" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341767 4894 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341776 4894 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341785 4894 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341795 4894 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341804 4894 flags.go:64] FLAG: --contention-profiling="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341813 4894 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341822 4894 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341832 4894 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341841 4894 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341899 4894 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341909 4894 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341918 4894 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341928 4894 flags.go:64] FLAG: --enable-load-reader="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341937 4894 flags.go:64] FLAG: --enable-server="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341946 4894 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341957 4894 flags.go:64] FLAG: --event-burst="100" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341966 4894 flags.go:64] FLAG: --event-qps="50" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341975 4894 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341985 4894 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.341994 4894 flags.go:64] FLAG: --eviction-hard="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342004 4894 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342013 4894 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342022 4894 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342033 4894 flags.go:64] FLAG: --eviction-soft="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342042 4894 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342051 4894 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342060 4894 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342069 4894 flags.go:64] FLAG: --experimental-mounter-path="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342079 4894 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342088 4894 flags.go:64] FLAG: --fail-swap-on="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342097 4894 flags.go:64] FLAG: --feature-gates="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342107 4894 flags.go:64] FLAG: --file-check-frequency="20s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342116 4894 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342126 4894 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342135 4894 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342145 4894 flags.go:64] FLAG: --healthz-port="10248" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342154 4894 flags.go:64] FLAG: --help="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342163 4894 flags.go:64] FLAG: --hostname-override="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342173 4894 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342182 4894 flags.go:64] FLAG: --http-check-frequency="20s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342192 4894 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342200 4894 flags.go:64] FLAG: --image-credential-provider-config="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342209 4894 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342218 4894 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342227 4894 flags.go:64] FLAG: --image-service-endpoint="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342236 4894 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342245 4894 flags.go:64] FLAG: --kube-api-burst="100" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342254 4894 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342263 4894 flags.go:64] FLAG: --kube-api-qps="50" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342272 4894 flags.go:64] FLAG: --kube-reserved="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342281 4894 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342290 4894 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342299 4894 flags.go:64] FLAG: --kubelet-cgroups="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342308 4894 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342317 4894 flags.go:64] FLAG: --lock-file="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342326 4894 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342336 4894 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342345 4894 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342358 4894 flags.go:64] FLAG: --log-json-split-stream="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342368 4894 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342378 4894 flags.go:64] FLAG: --log-text-split-stream="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342386 4894 flags.go:64] FLAG: --logging-format="text" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342395 4894 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342404 4894 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342413 4894 flags.go:64] FLAG: --manifest-url="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342422 4894 flags.go:64] FLAG: --manifest-url-header="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342434 4894 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342443 4894 flags.go:64] FLAG: --max-open-files="1000000" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342453 4894 flags.go:64] FLAG: --max-pods="110" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342462 4894 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342472 4894 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342481 4894 flags.go:64] FLAG: --memory-manager-policy="None" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342495 4894 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342504 4894 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342513 4894 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342523 4894 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342542 4894 flags.go:64] FLAG: --node-status-max-images="50" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342551 4894 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342560 4894 flags.go:64] FLAG: --oom-score-adj="-999" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342569 4894 flags.go:64] FLAG: --pod-cidr="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342579 4894 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342592 4894 flags.go:64] FLAG: --pod-manifest-path="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342601 4894 flags.go:64] FLAG: --pod-max-pids="-1" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342610 4894 flags.go:64] FLAG: --pods-per-core="0" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342620 4894 flags.go:64] FLAG: --port="10250" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342630 4894 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342639 4894 flags.go:64] FLAG: --provider-id="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342648 4894 flags.go:64] FLAG: --qos-reserved="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342656 4894 flags.go:64] FLAG: --read-only-port="10255" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342665 4894 flags.go:64] FLAG: --register-node="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342675 4894 flags.go:64] FLAG: --register-schedulable="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342684 4894 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342698 4894 flags.go:64] FLAG: --registry-burst="10" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342707 4894 flags.go:64] FLAG: --registry-qps="5" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342716 4894 flags.go:64] FLAG: --reserved-cpus="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342727 4894 flags.go:64] FLAG: --reserved-memory="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342744 4894 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342753 4894 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342763 4894 flags.go:64] FLAG: --rotate-certificates="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342771 4894 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342780 4894 flags.go:64] FLAG: --runonce="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342789 4894 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342799 4894 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342808 4894 flags.go:64] FLAG: --seccomp-default="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342817 4894 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342829 4894 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342839 4894 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342878 4894 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342887 4894 flags.go:64] FLAG: --storage-driver-password="root" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342896 4894 flags.go:64] FLAG: --storage-driver-secure="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342906 4894 flags.go:64] FLAG: --storage-driver-table="stats" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342915 4894 flags.go:64] FLAG: --storage-driver-user="root" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342923 4894 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342933 4894 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342942 4894 flags.go:64] FLAG: --system-cgroups="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342951 4894 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342966 4894 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342975 4894 flags.go:64] FLAG: --tls-cert-file="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342984 4894 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.342996 4894 flags.go:64] FLAG: --tls-min-version="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343004 4894 flags.go:64] FLAG: --tls-private-key-file="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343013 4894 flags.go:64] FLAG: --topology-manager-policy="none" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343049 4894 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343058 4894 flags.go:64] FLAG: --topology-manager-scope="container" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343067 4894 flags.go:64] FLAG: --v="2" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343078 4894 flags.go:64] FLAG: --version="false" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343089 4894 flags.go:64] FLAG: --vmodule="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343100 4894 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.343109 4894 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343317 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343328 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343340 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343351 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343360 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343369 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343378 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343387 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343399 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343408 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343417 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343424 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343433 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343441 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343449 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343457 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343465 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343473 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343481 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343489 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343500 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343508 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343516 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343524 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343532 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343540 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343549 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343557 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343565 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343574 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343582 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343590 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343599 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343607 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343615 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343626 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343636 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343644 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343662 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343671 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343683 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343692 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343700 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343708 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343716 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343726 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343737 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343747 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343756 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343765 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343773 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343782 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343792 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343800 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343808 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343817 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343825 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343833 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343841 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343873 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343882 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343890 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343898 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343906 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343914 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343921 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343930 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343938 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343946 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343955 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.343962 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.344218 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.356706 4894 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.356761 4894 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.356931 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.356957 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.356967 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.356977 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.356987 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.356995 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357004 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357012 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357020 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357030 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357041 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357050 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357058 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357066 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357074 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357082 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357089 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357097 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357105 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357113 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357120 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357128 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357135 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357144 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357151 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357159 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357167 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357174 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357182 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357190 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357197 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357208 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357218 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357227 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357238 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357246 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357254 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357261 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357269 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357277 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357284 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357292 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357302 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357312 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357320 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357328 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357337 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357345 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357353 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357362 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357371 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357380 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357388 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357395 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357403 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357411 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357419 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357426 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357434 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357441 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357449 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357457 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357464 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357472 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357480 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357488 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357495 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357504 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357512 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357519 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357528 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.357541 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357766 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357778 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357788 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357798 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357807 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357815 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357824 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357831 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357840 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357878 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357889 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357898 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357908 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357916 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357924 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357932 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357939 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357947 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357955 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357963 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357973 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357984 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.357993 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358003 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358011 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358019 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358027 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358035 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358043 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358051 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358059 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358067 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358074 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358082 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358092 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358100 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358107 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358115 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358123 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358130 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358138 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358146 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358153 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358161 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358169 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358177 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358184 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358192 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358199 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358207 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358215 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358222 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358230 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358237 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358245 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358255 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358265 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358275 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358283 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358292 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358300 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358308 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358316 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358324 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358331 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358339 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358347 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358355 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358362 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358370 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.358378 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.358392 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.358665 4894 server.go:940] "Client rotation is on, will bootstrap in background" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.362979 4894 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.363134 4894 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.364147 4894 server.go:997] "Starting client certificate rotation" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.364183 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.364809 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-12 00:30:20.852178185 +0000 UTC Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.364966 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 773h35m3.48722044s for next certificate rotation Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.371564 4894 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.373872 4894 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.387045 4894 log.go:25] "Validated CRI v1 runtime API" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.408251 4894 log.go:25] "Validated CRI v1 image API" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.410451 4894 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.413410 4894 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-10-18-50-57-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.413452 4894 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.444755 4894 manager.go:217] Machine: {Timestamp:2025-12-10 18:55:17.442647416 +0000 UTC m=+0.237495298 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:87dc1f21-a167-4b68-9d55-45806a47a7e3 BootID:c5920cce-fbb7-4806-aa5a-460529081e51 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:fb:20:0c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:fb:20:0c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:59:e4:41 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:8b:ae:c6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:33:82:25 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e1:8e:ea Speed:-1 Mtu:1496} {Name:eth10 MacAddress:22:3c:72:e0:f9:ad Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:ae:95:11:a5:36 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.445142 4894 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.445382 4894 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.445907 4894 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.446256 4894 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.446319 4894 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.446693 4894 topology_manager.go:138] "Creating topology manager with none policy" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.446721 4894 container_manager_linux.go:303] "Creating device plugin manager" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.447159 4894 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.447222 4894 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.447706 4894 state_mem.go:36] "Initialized new in-memory state store" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.447922 4894 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.449137 4894 kubelet.go:418] "Attempting to sync node with API server" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.449175 4894 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.449225 4894 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.449251 4894 kubelet.go:324] "Adding apiserver pod source" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.449274 4894 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.451444 4894 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.452047 4894 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.452756 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.452814 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.452935 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.452957 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.453325 4894 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454202 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454257 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454278 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454291 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454324 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454343 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454363 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454390 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454413 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454427 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454476 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454490 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.454761 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.455570 4894 server.go:1280] "Started kubelet" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.457233 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:17 crc systemd[1]: Started Kubernetes Kubelet. Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.457461 4894 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.457618 4894 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.459624 4894 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.462669 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.462733 4894 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.463470 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 19:05:52.156121902 +0000 UTC Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.464003 4894 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.464035 4894 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.464400 4894 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.464431 4894 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.465746 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="200ms" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.466798 4894 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.466991 4894 factory.go:55] Registering systemd factory Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.467307 4894 factory.go:221] Registration of the systemd container factory successfully Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.467762 4894 factory.go:153] Registering CRI-O factory Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.467207 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.467836 4894 factory.go:221] Registration of the crio container factory successfully Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.468052 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.470875 4894 factory.go:103] Registering Raw factory Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.471007 4894 manager.go:1196] Started watching for new ooms in manager Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.471749 4894 manager.go:319] Starting recovery of all containers Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.472340 4894 server.go:460] "Adding debug handlers to kubelet server" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.471241 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.201:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187fef8145ec9f12 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:55:17.455478546 +0000 UTC m=+0.250326368,LastTimestamp:2025-12-10 18:55:17.455478546 +0000 UTC m=+0.250326368,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479106 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479209 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479226 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479244 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479257 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479275 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479287 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479306 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479320 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479340 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479363 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479384 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.479400 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480247 4894 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480284 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480302 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480316 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480328 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480343 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480355 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480369 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480382 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480394 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480416 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480432 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480447 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480465 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480489 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480504 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480518 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480530 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480541 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480554 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480566 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480577 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480588 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480600 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480695 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480709 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480720 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480731 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480744 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480755 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480768 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480779 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480794 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480809 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480821 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480835 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480893 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480908 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480927 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480939 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480956 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480970 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.480987 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481004 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481021 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481037 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481054 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481066 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481078 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481098 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481110 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481128 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481140 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481151 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481164 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481177 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481189 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481203 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481216 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481227 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481238 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481249 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481261 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481271 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481292 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481304 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481315 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481328 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481338 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481349 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481360 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481372 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481383 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481396 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481445 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481485 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481497 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481509 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481521 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481533 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481551 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481564 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481575 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481590 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481602 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481614 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481626 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481638 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481649 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.481664 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485108 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485171 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485225 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485278 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485330 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485363 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485398 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.485436 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486074 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486123 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486168 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486202 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486241 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486269 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486299 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486336 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486363 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486398 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486431 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486458 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486493 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486523 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486563 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486591 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486621 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486660 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486687 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486724 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486752 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486780 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486818 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486936 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.486983 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487013 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487044 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487081 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487112 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487182 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487220 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487248 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487288 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487317 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487346 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487382 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487409 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487445 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487473 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487502 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487536 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487563 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487599 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487627 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487656 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487693 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487721 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487768 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487800 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.487830 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488091 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488124 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488155 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488194 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488220 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488257 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488285 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488315 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488352 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488380 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488419 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488449 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488476 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488512 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488543 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488583 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488612 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488637 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488674 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488701 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488734 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488760 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488790 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488825 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488886 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488924 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488950 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.488977 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.492038 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.492255 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.492557 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.492723 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.492792 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.492969 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493016 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493052 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493079 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493107 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493143 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493170 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493199 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493268 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493295 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493331 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493358 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493384 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493447 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493493 4894 reconstruct.go:97] "Volume reconstruction finished" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.493509 4894 reconciler.go:26] "Reconciler: start to sync state" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.501756 4894 manager.go:324] Recovery completed Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.512798 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.515049 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.515111 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.515129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.516236 4894 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.516255 4894 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.516277 4894 state_mem.go:36] "Initialized new in-memory state store" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.528002 4894 policy_none.go:49] "None policy: Start" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.528916 4894 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.528974 4894 state_mem.go:35] "Initializing new in-memory state store" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.537538 4894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.539578 4894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.539628 4894 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.539666 4894 kubelet.go:2335] "Starting kubelet main sync loop" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.539728 4894 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 10 18:55:17 crc kubenswrapper[4894]: W1210 18:55:17.540628 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.540697 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.565966 4894 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.594232 4894 manager.go:334] "Starting Device Plugin manager" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.594551 4894 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.594717 4894 server.go:79] "Starting device plugin registration server" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.595917 4894 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.596109 4894 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.596474 4894 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.596577 4894 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.596588 4894 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.604262 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.639980 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.640072 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.641124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.641161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.641171 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.641318 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.641708 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.641789 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.642326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.642368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.642380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.642493 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.642674 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.642719 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643642 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643768 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643813 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643783 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.643865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644644 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644810 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644859 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.644900 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.645511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.645550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.645565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.645932 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.645986 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.646344 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.646482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.646588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.647614 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.647723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.647808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.667492 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="400ms" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696236 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696286 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696322 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696414 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696497 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696558 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696586 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696660 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696718 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696751 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696772 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696812 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.696833 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.697813 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.697896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.697915 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.697948 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.698378 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.798531 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.798638 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.798697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.798743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.798784 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.798772 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799069 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799178 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799245 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799293 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799249 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799452 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799272 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799331 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799398 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799652 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799777 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799796 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799881 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.799975 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.800018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.899532 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.900983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.901035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.901053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.901094 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:17 crc kubenswrapper[4894]: E1210 18:55:17.901730 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.969562 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.985044 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:17 crc kubenswrapper[4894]: I1210 18:55:17.993130 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.009447 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-c33a3bb79d9f1c81c92470214543b05f0d8f50f1f28d09e8a1fc78b53acdd1b0 WatchSource:0}: Error finding container c33a3bb79d9f1c81c92470214543b05f0d8f50f1f28d09e8a1fc78b53acdd1b0: Status 404 returned error can't find the container with id c33a3bb79d9f1c81c92470214543b05f0d8f50f1f28d09e8a1fc78b53acdd1b0 Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.014060 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0a31a2c0b10a0e8225278f03d81651086d22d009665aac90b626bc85ef1730aa WatchSource:0}: Error finding container 0a31a2c0b10a0e8225278f03d81651086d22d009665aac90b626bc85ef1730aa: Status 404 returned error can't find the container with id 0a31a2c0b10a0e8225278f03d81651086d22d009665aac90b626bc85ef1730aa Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.015772 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.018276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.037642 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-c529aa39002150a4f927e91b6e6021844f1f09659678a7b8323b6926069f1061 WatchSource:0}: Error finding container c529aa39002150a4f927e91b6e6021844f1f09659678a7b8323b6926069f1061: Status 404 returned error can't find the container with id c529aa39002150a4f927e91b6e6021844f1f09659678a7b8323b6926069f1061 Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.040770 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-bfbc50d32f34f5f155f0eeb32e289e54c05f7967304ff70be1d1920547ba6dc4 WatchSource:0}: Error finding container bfbc50d32f34f5f155f0eeb32e289e54c05f7967304ff70be1d1920547ba6dc4: Status 404 returned error can't find the container with id bfbc50d32f34f5f155f0eeb32e289e54c05f7967304ff70be1d1920547ba6dc4 Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.045829 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-45451f0ffa19322c215b2ed8b38d80cf1616c2d356ed1efc5d98d967c3f653c3 WatchSource:0}: Error finding container 45451f0ffa19322c215b2ed8b38d80cf1616c2d356ed1efc5d98d967c3f653c3: Status 404 returned error can't find the container with id 45451f0ffa19322c215b2ed8b38d80cf1616c2d356ed1efc5d98d967c3f653c3 Dec 10 18:55:18 crc kubenswrapper[4894]: E1210 18:55:18.069463 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="800ms" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.301858 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.304354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.304401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.304415 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.304444 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:18 crc kubenswrapper[4894]: E1210 18:55:18.304966 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.460397 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.463747 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 22:39:49.266900939 +0000 UTC Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.549995 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0a31a2c0b10a0e8225278f03d81651086d22d009665aac90b626bc85ef1730aa"} Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.557214 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c33a3bb79d9f1c81c92470214543b05f0d8f50f1f28d09e8a1fc78b53acdd1b0"} Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.560180 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bfbc50d32f34f5f155f0eeb32e289e54c05f7967304ff70be1d1920547ba6dc4"} Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.563456 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45451f0ffa19322c215b2ed8b38d80cf1616c2d356ed1efc5d98d967c3f653c3"} Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.563631 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.565007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.565068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.565092 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.565666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c529aa39002150a4f927e91b6e6021844f1f09659678a7b8323b6926069f1061"} Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.565806 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.566928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.566999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:18 crc kubenswrapper[4894]: I1210 18:55:18.567023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.569119 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:18 crc kubenswrapper[4894]: E1210 18:55:18.569241 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.582306 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:18 crc kubenswrapper[4894]: E1210 18:55:18.582425 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:18 crc kubenswrapper[4894]: E1210 18:55:18.870693 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="1.6s" Dec 10 18:55:18 crc kubenswrapper[4894]: W1210 18:55:18.989342 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:18 crc kubenswrapper[4894]: E1210 18:55:18.989464 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4894]: W1210 18:55:19.048200 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4894]: E1210 18:55:19.048313 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.105510 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.107444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.107492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.107511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.107548 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:19 crc kubenswrapper[4894]: E1210 18:55:19.108096 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.459666 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.464245 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 20:27:19.795802302 +0000 UTC Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.464292 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 73h32m0.331513688s for next certificate rotation Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.572797 4894 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b" exitCode=0 Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.572908 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.573048 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.574691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.574724 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.574736 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.576439 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.576534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.576479 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.576589 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.576625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.578219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.578259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.578275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.579239 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="05de1c77ef9348557856937629d92ef297b1df9f1f5088495dc82bee78a91682" exitCode=0 Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.579317 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"05de1c77ef9348557856937629d92ef297b1df9f1f5088495dc82bee78a91682"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.579407 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.580462 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.580514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.580537 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.582446 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b" exitCode=0 Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.582556 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.582685 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.583758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.583787 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.583804 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.585327 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.586198 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.586228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.586266 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.586566 4894 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78" exitCode=0 Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.586630 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78"} Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.586762 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.587678 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.587869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.587901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4894]: I1210 18:55:19.587915 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4894]: W1210 18:55:20.316564 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:20 crc kubenswrapper[4894]: E1210 18:55:20.317064 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.390306 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.458209 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 18:55:20 crc kubenswrapper[4894]: E1210 18:55:20.471465 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="3.2s" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.591201 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc"} Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.591911 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.593116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.593155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.593167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.594959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99"} Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.595000 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d"} Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.596384 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ad957296dc59ff662f43ef9281358007bc20cbf238bfbcd4d279ee3a3a8cf5fe" exitCode=0 Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.596454 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ad957296dc59ff662f43ef9281358007bc20cbf238bfbcd4d279ee3a3a8cf5fe"} Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.599157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d"} Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.599224 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.599955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.599984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.599995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.708667 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.710081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.710118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.710127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4894]: I1210 18:55:20.710148 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.606682 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066"} Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.606758 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c"} Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.606781 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7"} Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.606798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4"} Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.606720 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.607637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.607674 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.607691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.610721 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f"} Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.610763 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.610871 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.610957 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612334 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4894]: I1210 18:55:21.612371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617323 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1f1d49683143a0573dcef072f12e72821265fcad39f733366bff084b222aa72d" exitCode=0 Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617429 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1f1d49683143a0573dcef072f12e72821265fcad39f733366bff084b222aa72d"} Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617483 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617536 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617544 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617608 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.617714 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619240 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619346 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4894]: I1210 18:55:22.619597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.525128 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.625521 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3e2f3dbf3c3ca45ced466e9506b9616d7a90ae0f135f70957daf77b2fee47286"} Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.625603 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bedaf28f139a42c66c14359ed98b88fb768677320df3716e4bfbd9e7286251a4"} Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.625630 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6b511f30df95cd098f87434d1b2c38933dc6e3763357256f87d0a54ce053aedd"} Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.625649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a5b5b629593dbea67a9a0bad2b12dae806d8c526d89250b7edd11005b265e571"} Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.625543 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.625729 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.627124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.627188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:23 crc kubenswrapper[4894]: I1210 18:55:23.627211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.365082 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.365284 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.366967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.367052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.367082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.396770 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.634754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ad46904f278522a83b4ac1e1b89abac0cea42c08790208964b8be7c20d1049af"} Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.634846 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.634906 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.634907 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.636126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.636220 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.636246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.636520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.636560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.636571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.912510 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.912810 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.914709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.914801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.914828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4894]: I1210 18:55:24.917649 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.638686 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.638732 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.640638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.640700 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.640720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.640646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.640930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.640955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:25 crc kubenswrapper[4894]: I1210 18:55:25.666375 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.641429 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.642391 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.642429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.642442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.799944 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.800163 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.801449 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.801500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:26 crc kubenswrapper[4894]: I1210 18:55:26.801514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.365375 4894 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.365467 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.503563 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.504294 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.509168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.509245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.509271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:27 crc kubenswrapper[4894]: E1210 18:55:27.604394 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.919159 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.919387 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.920644 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.920689 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:27 crc kubenswrapper[4894]: I1210 18:55:27.920703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:30 crc kubenswrapper[4894]: I1210 18:55:30.398158 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:30 crc kubenswrapper[4894]: I1210 18:55:30.398301 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:30 crc kubenswrapper[4894]: I1210 18:55:30.399522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:30 crc kubenswrapper[4894]: I1210 18:55:30.399584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:30 crc kubenswrapper[4894]: I1210 18:55:30.399608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:30 crc kubenswrapper[4894]: E1210 18:55:30.710884 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 10 18:55:31 crc kubenswrapper[4894]: W1210 18:55:31.088450 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:31 crc kubenswrapper[4894]: I1210 18:55:31.088575 4894 trace.go:236] Trace[1489563461]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:21.087) (total time: 10001ms): Dec 10 18:55:31 crc kubenswrapper[4894]: Trace[1489563461]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:55:31.088) Dec 10 18:55:31 crc kubenswrapper[4894]: Trace[1489563461]: [10.001483269s] [10.001483269s] END Dec 10 18:55:31 crc kubenswrapper[4894]: E1210 18:55:31.088605 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:31 crc kubenswrapper[4894]: W1210 18:55:31.435003 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:31 crc kubenswrapper[4894]: I1210 18:55:31.435202 4894 trace.go:236] Trace[11289836]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:21.433) (total time: 10001ms): Dec 10 18:55:31 crc kubenswrapper[4894]: Trace[11289836]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:55:31.434) Dec 10 18:55:31 crc kubenswrapper[4894]: Trace[11289836]: [10.001491895s] [10.001491895s] END Dec 10 18:55:31 crc kubenswrapper[4894]: E1210 18:55:31.435253 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:31 crc kubenswrapper[4894]: I1210 18:55:31.459148 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:31 crc kubenswrapper[4894]: W1210 18:55:31.558954 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:31 crc kubenswrapper[4894]: I1210 18:55:31.559078 4894 trace.go:236] Trace[1556381396]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:21.557) (total time: 10001ms): Dec 10 18:55:31 crc kubenswrapper[4894]: Trace[1556381396]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:55:31.558) Dec 10 18:55:31 crc kubenswrapper[4894]: Trace[1556381396]: [10.001521693s] [10.001521693s] END Dec 10 18:55:31 crc kubenswrapper[4894]: E1210 18:55:31.559108 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:31 crc kubenswrapper[4894]: E1210 18:55:31.868323 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187fef8145ec9f12 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:55:17.455478546 +0000 UTC m=+0.250326368,LastTimestamp:2025-12-10 18:55:17.455478546 +0000 UTC m=+0.250326368,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.525527 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.525617 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 18:55:33 crc kubenswrapper[4894]: E1210 18:55:33.672335 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.911785 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.913482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.913519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.913531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:33 crc kubenswrapper[4894]: I1210 18:55:33.913557 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:34 crc kubenswrapper[4894]: W1210 18:55:34.246315 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:34 crc kubenswrapper[4894]: I1210 18:55:34.246697 4894 trace.go:236] Trace[334424829]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:24.245) (total time: 10001ms): Dec 10 18:55:34 crc kubenswrapper[4894]: Trace[334424829]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:55:34.246) Dec 10 18:55:34 crc kubenswrapper[4894]: Trace[334424829]: [10.001382017s] [10.001382017s] END Dec 10 18:55:34 crc kubenswrapper[4894]: E1210 18:55:34.246918 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:35 crc kubenswrapper[4894]: I1210 18:55:35.704397 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 10 18:55:35 crc kubenswrapper[4894]: I1210 18:55:35.704691 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:35 crc kubenswrapper[4894]: I1210 18:55:35.706718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:35 crc kubenswrapper[4894]: I1210 18:55:35.706837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:35 crc kubenswrapper[4894]: I1210 18:55:35.706934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:35 crc kubenswrapper[4894]: I1210 18:55:35.725046 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 10 18:55:36 crc kubenswrapper[4894]: I1210 18:55:36.667105 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:36 crc kubenswrapper[4894]: I1210 18:55:36.668295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:36 crc kubenswrapper[4894]: I1210 18:55:36.668335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:36 crc kubenswrapper[4894]: I1210 18:55:36.668374 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:37 crc kubenswrapper[4894]: I1210 18:55:37.227537 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 18:55:37 crc kubenswrapper[4894]: I1210 18:55:37.227604 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 18:55:37 crc kubenswrapper[4894]: I1210 18:55:37.366860 4894 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 18:55:37 crc kubenswrapper[4894]: I1210 18:55:37.366942 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 18:55:37 crc kubenswrapper[4894]: E1210 18:55:37.604528 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.532955 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.533207 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.534720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.534779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.534803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.540745 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.671172 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.671228 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.672452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.672516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:38 crc kubenswrapper[4894]: I1210 18:55:38.672536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:40 crc kubenswrapper[4894]: I1210 18:55:40.711835 4894 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.460518 4894 apiserver.go:52] "Watching apiserver" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.464494 4894 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.464899 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.465385 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.465495 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.465669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.465686 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:41 crc kubenswrapper[4894]: E1210 18:55:41.465737 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:41 crc kubenswrapper[4894]: E1210 18:55:41.466250 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.466660 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.466731 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:41 crc kubenswrapper[4894]: E1210 18:55:41.467743 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.468689 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.468722 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.468790 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.469073 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.469207 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.469270 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.469295 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.470032 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.470163 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.502469 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.523780 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.535836 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.552926 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.565357 4894 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.570288 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.582092 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:41 crc kubenswrapper[4894]: I1210 18:55:41.593929 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.193028 4894 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.194096 4894 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.194487 4894 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.195080 4894 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.195731 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.227944 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34664->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.228068 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34664->192.168.126.11:17697: read: connection reset by peer" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.227944 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59638->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.228218 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59638->192.168.126.11:17697: read: connection reset by peer" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.228653 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.228732 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295347 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295431 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295470 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295517 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295547 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295600 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295628 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295659 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295683 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295708 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295735 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295763 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295815 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.295895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296077 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296106 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296138 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296165 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296188 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296213 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296238 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296264 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296293 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296284 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296334 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296322 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296491 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296536 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296574 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296596 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296618 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296637 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296656 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296681 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296699 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296721 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296717 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296747 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296768 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296812 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296833 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296812 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296814 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296829 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296819 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.296991 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297032 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297054 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297078 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.297107 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.797075064 +0000 UTC m=+25.591923066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297110 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297113 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297111 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297159 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297176 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297233 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297255 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297308 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297343 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297365 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297375 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297464 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297463 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297491 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297492 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297520 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297562 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297586 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297606 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297617 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297623 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297682 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297692 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297716 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297721 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297732 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297756 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297783 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297838 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297884 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297917 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297943 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297970 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297993 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298017 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298034 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298042 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298102 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298113 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298173 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298187 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298264 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298319 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298381 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298431 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298477 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298525 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298569 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298600 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298624 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298648 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298672 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298694 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298716 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298739 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298764 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298793 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298832 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298889 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298923 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298954 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298990 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299030 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299061 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299093 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299128 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299176 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299198 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299238 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299271 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299302 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299329 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299353 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299376 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299428 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299450 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299474 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299497 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299519 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299544 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299598 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299622 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299645 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299682 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299704 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299737 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299783 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299805 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299829 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299878 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299904 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299927 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299950 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299975 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299997 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300095 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300123 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300150 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300173 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301331 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301367 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301397 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301425 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301459 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301491 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301520 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301552 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301583 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301615 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301651 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301684 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301718 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301754 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301786 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301816 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301883 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301916 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301944 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301977 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302042 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302074 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302133 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302166 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302204 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302236 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302268 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298386 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298633 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298775 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298787 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.298893 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299192 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299235 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299321 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299354 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299387 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.299639 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300087 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.297136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300233 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300439 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300450 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300800 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300901 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.300962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301406 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.301959 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302283 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302361 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302408 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302513 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302673 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.302986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303095 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303106 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303355 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303404 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303441 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303489 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303735 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303767 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303773 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303905 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.303982 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304013 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304047 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304113 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304144 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304177 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304207 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304237 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304269 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304304 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304339 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304374 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304409 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304444 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304479 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304512 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304551 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304588 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304620 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304656 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304691 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304723 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304912 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304947 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.304987 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305101 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305139 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305301 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305467 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305490 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305511 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305533 4894 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305556 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305578 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305597 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305822 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305617 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305971 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305947 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.305980 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306051 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306075 4894 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306131 4894 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306151 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306173 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306193 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306213 4894 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306348 4894 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306371 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306484 4894 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306505 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306523 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306546 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306565 4894 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306584 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306607 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306628 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306648 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306677 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306046 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306027 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306076 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306410 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306742 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.306883 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.307072 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.307535 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.307567 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.307790 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.308017 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.308925 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309140 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309561 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309578 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309773 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309822 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309992 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.309841 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.310348 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.310349 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.310863 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.311174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.311551 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.311576 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.311663 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.311606 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.311832 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.313272 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.313295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.313225 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.313339 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.313687 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.314053 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.314516 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.314631 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.314765 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.314920 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.314982 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.316321 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.316493 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.316371 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.316883 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.317196 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.317264 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.317290 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.317398 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.315211 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.318117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.315257 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.318192 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.317738 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.317800 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.318449 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.318555 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.818527221 +0000 UTC m=+25.613375223 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319160 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319208 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319453 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319549 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319787 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320049 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320837 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.319177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320953 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321013 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321052 4894 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321088 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321284 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321418 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321416 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321427 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321593 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321634 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.321791 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320841 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320315 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.320448 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.322901 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.822880276 +0000 UTC m=+25.617728068 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320494 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.320719 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.324913 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.325667 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.325879 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.330157 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.330199 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.330212 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.330273 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.830254725 +0000 UTC m=+25.625102517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.333743 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.334094 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.334460 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.335478 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336217 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336508 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336552 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336561 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336638 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.336780 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338121 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338137 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338236 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338247 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338536 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338612 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.338804 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.339103 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.339243 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.339471 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.341062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.343978 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.343978 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.345044 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.345048 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.345103 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.345212 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.345250 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.345281 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.345301 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.345278 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.345370 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.845348433 +0000 UTC m=+25.640196225 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.345721 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.346008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.346257 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.346361 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.346610 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347016 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347513 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347734 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347651 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347819 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347826 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347861 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.347879 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.348088 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.348296 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.348331 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.348540 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.348966 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.349470 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.350877 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.351014 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.351105 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.351169 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.351248 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.351890 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.353705 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.360807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.369566 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.383986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.386837 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408034 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408160 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408173 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408185 4894 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408198 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408206 4894 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408215 4894 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408226 4894 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408236 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408246 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408256 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408268 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408276 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408285 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408294 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408306 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408315 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408326 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408336 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408344 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408353 4894 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408364 4894 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408376 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408385 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408394 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408402 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408412 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408421 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408429 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408438 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408449 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408457 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408465 4894 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408476 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408484 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408493 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408502 4894 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408512 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408522 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408532 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408540 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408551 4894 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408559 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408568 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408580 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408588 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408597 4894 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408607 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408621 4894 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408632 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408644 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408653 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408665 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408653 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408731 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408674 4894 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408828 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408865 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408878 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408895 4894 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408904 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408940 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408949 4894 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408960 4894 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408969 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408978 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408986 4894 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.408998 4894 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409007 4894 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409016 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409024 4894 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409035 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409042 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409051 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409065 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409075 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409083 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409091 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409102 4894 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409110 4894 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409119 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409127 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409138 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409147 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409155 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409163 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409177 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409185 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409194 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409204 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409213 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409221 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409229 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409240 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409250 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409257 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409268 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409278 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409286 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409294 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409304 4894 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409312 4894 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409320 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409328 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409339 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409347 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409355 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409363 4894 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409373 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409384 4894 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409393 4894 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409402 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409413 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409423 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409431 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409442 4894 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409451 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409459 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409467 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409478 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409486 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409494 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409502 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409513 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409521 4894 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409529 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409540 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409548 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409558 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409567 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409577 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409586 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409594 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409602 4894 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409613 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409625 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409637 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409647 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409658 4894 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409667 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409675 4894 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409686 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409694 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409703 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409712 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409722 4894 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409731 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409739 4894 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409748 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409758 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409767 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409774 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409785 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409793 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409802 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409810 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409820 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409829 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409837 4894 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409864 4894 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409880 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409888 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409897 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409905 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409916 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409924 4894 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409933 4894 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409945 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.409954 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.413793 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.420649 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:42 crc kubenswrapper[4894]: W1210 18:55:42.427010 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-41f97bcfea2a04652b54059c8ccf4fc215ec5ca09c379977945edb5d1edfb370 WatchSource:0}: Error finding container 41f97bcfea2a04652b54059c8ccf4fc215ec5ca09c379977945edb5d1edfb370: Status 404 returned error can't find the container with id 41f97bcfea2a04652b54059c8ccf4fc215ec5ca09c379977945edb5d1edfb370 Dec 10 18:55:42 crc kubenswrapper[4894]: W1210 18:55:42.431133 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3de3d7cd152635629588131b83d6236c072456a8824029305ea89515a86d8e96 WatchSource:0}: Error finding container 3de3d7cd152635629588131b83d6236c072456a8824029305ea89515a86d8e96: Status 404 returned error can't find the container with id 3de3d7cd152635629588131b83d6236c072456a8824029305ea89515a86d8e96 Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.539935 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.540044 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.682974 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3de3d7cd152635629588131b83d6236c072456a8824029305ea89515a86d8e96"} Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.684735 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257"} Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.684778 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1"} Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.684789 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"41f97bcfea2a04652b54059c8ccf4fc215ec5ca09c379977945edb5d1edfb370"} Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.686446 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.688092 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066" exitCode=255 Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.688126 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066"} Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.697254 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.697325 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.697507 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.697668 4894 scope.go:117] "RemoveContainer" containerID="b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066" Dec 10 18:55:42 crc kubenswrapper[4894]: W1210 18:55:42.708210 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-19b1dd6ec22e2bd247ad74c7f0ad09224bd00f1e8a08c9a83e430ac1e7e624f9 WatchSource:0}: Error finding container 19b1dd6ec22e2bd247ad74c7f0ad09224bd00f1e8a08c9a83e430ac1e7e624f9: Status 404 returned error can't find the container with id 19b1dd6ec22e2bd247ad74c7f0ad09224bd00f1e8a08c9a83e430ac1e7e624f9 Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.709608 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.737395 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.752183 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.789168 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.812937 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.813130 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:43.813101796 +0000 UTC m=+26.607949588 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.815499 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.830346 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.852029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.864671 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.878271 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.894149 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.906276 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.913619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.913673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.913692 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.913713 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913785 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913802 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913830 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:43.913817133 +0000 UTC m=+26.708664935 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913868 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913894 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913904 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913917 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913924 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913906 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.913872 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:43.913854874 +0000 UTC m=+26.708702666 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.914008 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:43.914001438 +0000 UTC m=+26.708849230 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: E1210 18:55:42.914018 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:43.914013949 +0000 UTC m=+26.708861741 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4894]: I1210 18:55:42.917674 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.540760 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.540862 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.540950 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.541077 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.545586 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.546723 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.548653 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.549801 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.551505 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.552351 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.553203 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.554661 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.555764 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.557063 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.557937 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.559559 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.560274 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.561114 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.562806 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.563617 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.565230 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.565869 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.566638 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.567689 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.568164 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.569200 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.569634 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.570642 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.571210 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.571783 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.572808 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.573311 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.574531 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.575010 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.576001 4894 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.576116 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.577936 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.578880 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.579405 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.580786 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.581549 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.582383 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.582984 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.584050 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.584506 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.585450 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.586298 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.587158 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.588529 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.589276 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.590548 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.591803 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.593230 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.593974 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.594707 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.596104 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.597169 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.598438 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.693874 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.695648 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09"} Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.696083 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.697432 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04"} Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.697564 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"19b1dd6ec22e2bd247ad74c7f0ad09224bd00f1e8a08c9a83e430ac1e7e624f9"} Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.729467 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.764811 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.797262 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.813253 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.823416 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.823608 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:45.823582698 +0000 UTC m=+28.618430500 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.832783 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.853379 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.864273 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.874147 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.885596 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.901675 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.930960 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.939558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.939635 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.939683 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.939708 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939763 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939809 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939867 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:45.939830217 +0000 UTC m=+28.734678029 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939809 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939898 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:45.939883038 +0000 UTC m=+28.734730840 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939920 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939954 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939968 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.940019 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:45.940001222 +0000 UTC m=+28.734849094 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.939902 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.940048 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:43 crc kubenswrapper[4894]: E1210 18:55:43.940111 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:45.940102095 +0000 UTC m=+28.734950017 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.953162 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.967254 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4894]: I1210 18:55:43.981173 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.374790 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.383461 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.389573 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.413546 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.448685 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.462951 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.480949 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.495395 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.509375 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.529237 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.540496 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:44 crc kubenswrapper[4894]: E1210 18:55:44.540614 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.553704 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.569037 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.593524 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.616898 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.634229 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.648727 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.662638 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4894]: I1210 18:55:44.677725 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.541172 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.541180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.541441 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.541408 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.704833 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b"} Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.734086 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.753496 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.771768 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.789434 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.812283 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.832433 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.854050 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.857894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.858181 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:49.858140604 +0000 UTC m=+32.652988436 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.874140 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.959220 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.959356 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.959415 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:45 crc kubenswrapper[4894]: I1210 18:55:45.959472 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.959603 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.959690 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:49.959662855 +0000 UTC m=+32.754510697 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960104 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960149 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960203 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960225 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960172 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:49.960156501 +0000 UTC m=+32.755004303 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960314 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:49.960284195 +0000 UTC m=+32.755132027 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960416 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960501 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960530 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:45 crc kubenswrapper[4894]: E1210 18:55:45.960631 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:49.960598755 +0000 UTC m=+32.755446597 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:46 crc kubenswrapper[4894]: I1210 18:55:46.540515 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:46 crc kubenswrapper[4894]: E1210 18:55:46.540733 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.540739 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.540739 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:47 crc kubenswrapper[4894]: E1210 18:55:47.540991 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:47 crc kubenswrapper[4894]: E1210 18:55:47.541170 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.557110 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.572011 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.588325 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.609450 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.634179 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.651434 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.665963 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4894]: I1210 18:55:47.680956 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.540154 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.540272 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.596197 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.597782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.597813 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.597822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.597899 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.604398 4894 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.604706 4894 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.605683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.605714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.605725 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.605741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.605754 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.621002 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:48Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.623927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.623949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.623957 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.623969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.623978 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.635694 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:48Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.638703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.638732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.638761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.638779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.638788 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.650830 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:48Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.653686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.653726 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.653738 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.653755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.653768 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.663542 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:48Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.666476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.666512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.666521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.666534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.666546 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.691993 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:48Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:48 crc kubenswrapper[4894]: E1210 18:55:48.692109 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.693947 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.693990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.694001 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.694014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.694023 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.795798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.795828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.795837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.795861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.795870 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.901162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.901229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.901244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.901263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4894]: I1210 18:55:48.901280 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.003812 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.003876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.003886 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.003903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.003913 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.105753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.105805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.105817 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.105833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.105858 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.207713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.207746 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.207754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.207769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.207778 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.241333 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-msfl4"] Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.241627 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-wnfrc"] Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.241800 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-q8rg7"] Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.242471 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.242903 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.243289 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.244897 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-vbjj4"] Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.245205 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.246315 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.246732 4894 reflector.go:561] object-"openshift-dns"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.246765 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.247086 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.247258 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.247478 4894 reflector.go:561] object-"openshift-multus"/"default-cni-sysctl-allowlist": failed to list *v1.ConfigMap: configmaps "default-cni-sysctl-allowlist" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.247503 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"default-cni-sysctl-allowlist\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.247888 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.248025 4894 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.248047 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.248120 4894 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.248140 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.248193 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.248355 4894 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.248378 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.248599 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.249333 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.249662 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.250416 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.250664 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.259667 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.268712 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.281865 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.293231 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.304414 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.309763 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.309801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.309811 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.309825 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.309835 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.318027 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.341769 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.356375 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.371165 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.385241 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390721 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-k8s-cni-cncf-io\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390760 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-netns\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390780 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-etc-kubernetes\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390798 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-tuning-conf-dir\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390816 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-kubelet\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390867 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-hostroot\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390885 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64fjr\" (UniqueName: \"kubernetes.io/projected/d9392e0e-ca87-437a-8ba1-9c5d34a50660-kube-api-access-64fjr\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.390969 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-socket-dir-parent\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391014 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szhgj\" (UniqueName: \"kubernetes.io/projected/7c842b50-bc0f-47a6-97aa-67c9562005aa-kube-api-access-szhgj\") pod \"node-resolver-wnfrc\" (UID: \"7c842b50-bc0f-47a6-97aa-67c9562005aa\") " pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391040 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bb325131-20c4-4791-9bbc-45ae36443184-cni-binary-copy\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-cni-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391083 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-cnibin\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391103 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-cnibin\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391129 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3fd55194-d263-4547-9b8f-934ef8050b59-rootfs\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-system-cni-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-os-release\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391325 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bb325131-20c4-4791-9bbc-45ae36443184-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391367 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3fd55194-d263-4547-9b8f-934ef8050b59-mcd-auth-proxy-config\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391444 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d9392e0e-ca87-437a-8ba1-9c5d34a50660-cni-binary-copy\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391525 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-os-release\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-cni-bin\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391582 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3fd55194-d263-4547-9b8f-934ef8050b59-proxy-tls\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391601 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-conf-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391647 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-multus-certs\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391688 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c842b50-bc0f-47a6-97aa-67c9562005aa-hosts-file\") pod \"node-resolver-wnfrc\" (UID: \"7c842b50-bc0f-47a6-97aa-67c9562005aa\") " pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-system-cni-dir\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391737 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvr6g\" (UniqueName: \"kubernetes.io/projected/bb325131-20c4-4791-9bbc-45ae36443184-kube-api-access-pvr6g\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391759 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ftsp\" (UniqueName: \"kubernetes.io/projected/3fd55194-d263-4547-9b8f-934ef8050b59-kube-api-access-5ftsp\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-daemon-config\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.391839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-cni-multus\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.400327 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.412158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.412194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.412203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.412217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.412227 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.413048 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.424002 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.434289 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.446429 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.456731 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.472886 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.482821 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493044 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3fd55194-d263-4547-9b8f-934ef8050b59-proxy-tls\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493086 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-conf-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-multus-certs\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493139 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c842b50-bc0f-47a6-97aa-67c9562005aa-hosts-file\") pod \"node-resolver-wnfrc\" (UID: \"7c842b50-bc0f-47a6-97aa-67c9562005aa\") " pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-system-cni-dir\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493184 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvr6g\" (UniqueName: \"kubernetes.io/projected/bb325131-20c4-4791-9bbc-45ae36443184-kube-api-access-pvr6g\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493206 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ftsp\" (UniqueName: \"kubernetes.io/projected/3fd55194-d263-4547-9b8f-934ef8050b59-kube-api-access-5ftsp\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-daemon-config\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493229 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-multus-certs\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493242 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-conf-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493259 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-cni-multus\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493311 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-cni-multus\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-system-cni-dir\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c842b50-bc0f-47a6-97aa-67c9562005aa-hosts-file\") pod \"node-resolver-wnfrc\" (UID: \"7c842b50-bc0f-47a6-97aa-67c9562005aa\") " pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493638 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-k8s-cni-cncf-io\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493674 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-netns\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493702 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-tuning-conf-dir\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-kubelet\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493739 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-netns\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493745 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-hostroot\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493770 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-hostroot\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493767 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-run-k8s-cni-cncf-io\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-etc-kubernetes\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493812 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-kubelet\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493821 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64fjr\" (UniqueName: \"kubernetes.io/projected/d9392e0e-ca87-437a-8ba1-9c5d34a50660-kube-api-access-64fjr\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493856 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-socket-dir-parent\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493864 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-etc-kubernetes\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493877 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szhgj\" (UniqueName: \"kubernetes.io/projected/7c842b50-bc0f-47a6-97aa-67c9562005aa-kube-api-access-szhgj\") pod \"node-resolver-wnfrc\" (UID: \"7c842b50-bc0f-47a6-97aa-67c9562005aa\") " pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493896 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-daemon-config\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-cni-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493928 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-socket-dir-parent\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493942 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-multus-cni-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493949 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-cnibin\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493971 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bb325131-20c4-4791-9bbc-45ae36443184-cni-binary-copy\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.493988 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-system-cni-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494004 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-cnibin\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494020 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3fd55194-d263-4547-9b8f-934ef8050b59-rootfs\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494041 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3fd55194-d263-4547-9b8f-934ef8050b59-mcd-auth-proxy-config\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d9392e0e-ca87-437a-8ba1-9c5d34a50660-cni-binary-copy\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-os-release\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494080 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-cnibin\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494112 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3fd55194-d263-4547-9b8f-934ef8050b59-rootfs\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494150 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-os-release\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494084 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bb325131-20c4-4791-9bbc-45ae36443184-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494085 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-cnibin\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494197 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-os-release\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-system-cni-dir\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494173 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-os-release\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-cni-bin\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d9392e0e-ca87-437a-8ba1-9c5d34a50660-host-var-lib-cni-bin\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494619 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d9392e0e-ca87-437a-8ba1-9c5d34a50660-cni-binary-copy\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494666 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bb325131-20c4-4791-9bbc-45ae36443184-cni-binary-copy\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494707 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bb325131-20c4-4791-9bbc-45ae36443184-tuning-conf-dir\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.494836 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3fd55194-d263-4547-9b8f-934ef8050b59-mcd-auth-proxy-config\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.495797 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.500087 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3fd55194-d263-4547-9b8f-934ef8050b59-proxy-tls\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.506988 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.509456 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64fjr\" (UniqueName: \"kubernetes.io/projected/d9392e0e-ca87-437a-8ba1-9c5d34a50660-kube-api-access-64fjr\") pod \"multus-msfl4\" (UID: \"d9392e0e-ca87-437a-8ba1-9c5d34a50660\") " pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.513384 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvr6g\" (UniqueName: \"kubernetes.io/projected/bb325131-20c4-4791-9bbc-45ae36443184-kube-api-access-pvr6g\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.514256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.514280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.514288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.514300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.514310 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.518915 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.519309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ftsp\" (UniqueName: \"kubernetes.io/projected/3fd55194-d263-4547-9b8f-934ef8050b59-kube-api-access-5ftsp\") pod \"machine-config-daemon-vbjj4\" (UID: \"3fd55194-d263-4547-9b8f-934ef8050b59\") " pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.531094 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.540494 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.540494 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.540614 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.540700 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.541709 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.563375 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-msfl4" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.576097 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.577014 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9392e0e_ca87_437a_8ba1_9c5d34a50660.slice/crio-509d1b13989842819fbefec85ecbfbec74f76fab16ea1573304c1572941ec092 WatchSource:0}: Error finding container 509d1b13989842819fbefec85ecbfbec74f76fab16ea1573304c1572941ec092: Status 404 returned error can't find the container with id 509d1b13989842819fbefec85ecbfbec74f76fab16ea1573304c1572941ec092 Dec 10 18:55:49 crc kubenswrapper[4894]: W1210 18:55:49.585463 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fd55194_d263_4547_9b8f_934ef8050b59.slice/crio-6c1ef30a2e0020b24d15d5af6ba0a952df17c30ee6ebda7533e82ae283c75ba2 WatchSource:0}: Error finding container 6c1ef30a2e0020b24d15d5af6ba0a952df17c30ee6ebda7533e82ae283c75ba2: Status 404 returned error can't find the container with id 6c1ef30a2e0020b24d15d5af6ba0a952df17c30ee6ebda7533e82ae283c75ba2 Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.603115 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jqwjd"] Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.604104 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.606736 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.606949 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.606751 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.607166 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.607348 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.608304 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.608648 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.616389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.616583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.616646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.616710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.616767 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.618862 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.634677 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.647740 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.662877 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.675880 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.689017 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696498 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-ovn\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696546 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-script-lib\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696582 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-systemd-units\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696606 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696627 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-ovn-kubernetes\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696648 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-var-lib-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696670 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovn-node-metrics-cert\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.696841 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-slash\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697003 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-etc-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697056 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-bin\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-config\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697142 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-env-overrides\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697174 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-kubelet\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-node-log\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnqbs\" (UniqueName: \"kubernetes.io/projected/8de31cb9-e20f-4578-b0ac-b03df0bc133c-kube-api-access-bnqbs\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-netd\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697367 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-netns\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697399 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697468 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-systemd\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.697503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-log-socket\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.705531 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.715221 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"6c1ef30a2e0020b24d15d5af6ba0a952df17c30ee6ebda7533e82ae283c75ba2"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.716401 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerStarted","Data":"509d1b13989842819fbefec85ecbfbec74f76fab16ea1573304c1572941ec092"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.719292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.719360 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.719378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.719402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.719420 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.723448 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.737323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.746970 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.769138 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.785275 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-var-lib-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798055 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovn-node-metrics-cert\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-slash\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-etc-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798159 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-bin\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798174 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-var-lib-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798236 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-slash\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798284 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-etc-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798314 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-bin\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798184 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-config\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798358 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-env-overrides\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798381 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-kubelet\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-node-log\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798434 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnqbs\" (UniqueName: \"kubernetes.io/projected/8de31cb9-e20f-4578-b0ac-b03df0bc133c-kube-api-access-bnqbs\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798463 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-netd\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-netns\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798519 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-systemd\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-log-socket\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798589 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-ovn\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-node-log\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-netd\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798662 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-netns\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798657 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-kubelet\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798697 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-systemd\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798734 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-log-socket\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798749 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-script-lib\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798780 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-ovn\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798937 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-config\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.798994 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-systemd-units\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799027 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-systemd-units\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799052 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-ovn-kubernetes\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799144 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-ovn-kubernetes\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799149 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-openvswitch\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799187 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-env-overrides\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.799350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-script-lib\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.800123 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.801517 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovn-node-metrics-cert\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.816570 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnqbs\" (UniqueName: \"kubernetes.io/projected/8de31cb9-e20f-4578-b0ac-b03df0bc133c-kube-api-access-bnqbs\") pod \"ovnkube-node-jqwjd\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.822053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.822085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.822094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.822110 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.822119 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.900206 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:49 crc kubenswrapper[4894]: E1210 18:55:49.900405 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:57.900390241 +0000 UTC m=+40.695238033 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.924290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.924324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.924335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.924350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4894]: I1210 18:55:49.924363 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.001491 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.001531 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.001568 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.001587 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.001652 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.001696 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:58.001683555 +0000 UTC m=+40.796531347 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.001999 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002024 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:58.002017465 +0000 UTC m=+40.796865257 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002085 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002098 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002107 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002127 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:58.002121378 +0000 UTC m=+40.796969170 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002162 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002170 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002176 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.002193 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:58.00218752 +0000 UTC m=+40.797035312 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.019029 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.026711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.026776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.026786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.026802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.026813 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: W1210 18:55:50.030039 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8de31cb9_e20f_4578_b0ac_b03df0bc133c.slice/crio-760360a48f31b9ed7a20ab001b50f3b54e7cfe78684b3dc94edc35a1d3eac9f4 WatchSource:0}: Error finding container 760360a48f31b9ed7a20ab001b50f3b54e7cfe78684b3dc94edc35a1d3eac9f4: Status 404 returned error can't find the container with id 760360a48f31b9ed7a20ab001b50f3b54e7cfe78684b3dc94edc35a1d3eac9f4 Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.089902 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.095073 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bb325131-20c4-4791-9bbc-45ae36443184-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-q8rg7\" (UID: \"bb325131-20c4-4791-9bbc-45ae36443184\") " pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.129627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.129651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.129658 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.129671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.129680 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.200389 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.206704 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" Dec 10 18:55:50 crc kubenswrapper[4894]: W1210 18:55:50.216900 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb325131_20c4_4791_9bbc_45ae36443184.slice/crio-53101d9941d2532ff5fec3da7227de7f80326c928c953b6baf6c51dda17471e5 WatchSource:0}: Error finding container 53101d9941d2532ff5fec3da7227de7f80326c928c953b6baf6c51dda17471e5: Status 404 returned error can't find the container with id 53101d9941d2532ff5fec3da7227de7f80326c928c953b6baf6c51dda17471e5 Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.232478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.232523 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.232536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.232552 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.232562 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.335032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.335086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.335095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.335109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.335120 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.341962 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.384816 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.395599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szhgj\" (UniqueName: \"kubernetes.io/projected/7c842b50-bc0f-47a6-97aa-67c9562005aa-kube-api-access-szhgj\") pod \"node-resolver-wnfrc\" (UID: \"7c842b50-bc0f-47a6-97aa-67c9562005aa\") " pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.438925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.438958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.438972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.438989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.439001 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.539958 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:50 crc kubenswrapper[4894]: E1210 18:55:50.540116 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.540908 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.540942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.540952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.540965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.540974 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.633487 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.643263 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wnfrc" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.643761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.643931 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.643949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.643972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.643990 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.721715 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wnfrc" event={"ID":"7c842b50-bc0f-47a6-97aa-67c9562005aa","Type":"ContainerStarted","Data":"1f0026fb8fc2e56829e1bcd96d11c0446a87019e0312ee96920ce6ee225f320a"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.725679 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4" exitCode=0 Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.725726 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.725742 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"760360a48f31b9ed7a20ab001b50f3b54e7cfe78684b3dc94edc35a1d3eac9f4"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.730657 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerStarted","Data":"a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.731704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerStarted","Data":"e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.731729 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerStarted","Data":"53101d9941d2532ff5fec3da7227de7f80326c928c953b6baf6c51dda17471e5"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.734006 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.734029 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.746375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.746406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.746416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.746431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.746442 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.746448 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.762559 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.774444 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.788936 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.805501 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.818005 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.833602 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.849511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.849688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.849699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.849714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.849723 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.850357 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.871264 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.889778 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.903819 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.921685 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.932746 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.946177 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.953346 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.953376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.953387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.953403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.953412 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.962932 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.979333 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4894]: I1210 18:55:50.993144 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.005590 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.016784 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.027585 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.044279 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.056213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.056254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.056265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.056281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.056292 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.057882 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.069452 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.081070 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.092290 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.105255 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.159610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.159642 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.159652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.159665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.159673 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.261352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.261394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.261409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.261426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.261447 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.363564 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.363876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.363987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.364133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.364238 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.466452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.466524 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.466550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.466577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.466599 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.540353 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.540409 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:51 crc kubenswrapper[4894]: E1210 18:55:51.540874 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:51 crc kubenswrapper[4894]: E1210 18:55:51.540749 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.568870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.569102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.569178 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.569284 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.569379 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.671970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.672005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.672016 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.672032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.672043 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.739588 4894 generic.go:334] "Generic (PLEG): container finished" podID="bb325131-20c4-4791-9bbc-45ae36443184" containerID="e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536" exitCode=0 Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.739684 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerDied","Data":"e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.743428 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.743462 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.743475 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.743487 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.743497 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.743507 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.744784 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wnfrc" event={"ID":"7c842b50-bc0f-47a6-97aa-67c9562005aa","Type":"ContainerStarted","Data":"0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.767182 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.774442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.774470 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.774479 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.774496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.774508 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.778485 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.799611 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.819125 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.830660 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.842680 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.853379 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.867713 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.876798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.876833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.876856 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.876875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.876886 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.879136 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.896897 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.907389 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.922720 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.933480 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.946496 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.959387 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.972468 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.978688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.978802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.978997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.979102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.979195 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.983222 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4894]: I1210 18:55:51.993771 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.009868 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.020662 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.035233 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.049430 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.057694 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.075445 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.081932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.082292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.082310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.082329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.082341 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.089429 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.101051 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.185498 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.185605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.185617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.185637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.185649 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.287833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.287923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.287942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.287965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.287984 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.390706 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.390735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.390746 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.390758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.390769 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.492927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.492967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.492975 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.492991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.493000 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.540687 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:52 crc kubenswrapper[4894]: E1210 18:55:52.540826 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.573978 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-n6rc2"] Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.574580 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.577786 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.577977 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.578117 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.578120 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.595012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.595076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.595100 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.595128 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.595155 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.598361 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.611982 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.636223 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.653448 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.669241 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.683032 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.697871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.697914 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.697927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.697945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.697956 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.701390 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.716560 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.730669 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.732163 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f663d56-37e2-4a3f-9bb1-24e31b011470-host\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.732216 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czrsx\" (UniqueName: \"kubernetes.io/projected/6f663d56-37e2-4a3f-9bb1-24e31b011470-kube-api-access-czrsx\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.732262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6f663d56-37e2-4a3f-9bb1-24e31b011470-serviceca\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.747703 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.751398 4894 generic.go:334] "Generic (PLEG): container finished" podID="bb325131-20c4-4791-9bbc-45ae36443184" containerID="14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180" exitCode=0 Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.751496 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerDied","Data":"14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.768879 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.792195 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.800233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.800290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.800309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.800333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.800350 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.808646 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.823491 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.833046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f663d56-37e2-4a3f-9bb1-24e31b011470-host\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.833090 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czrsx\" (UniqueName: \"kubernetes.io/projected/6f663d56-37e2-4a3f-9bb1-24e31b011470-kube-api-access-czrsx\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.833125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6f663d56-37e2-4a3f-9bb1-24e31b011470-serviceca\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.833721 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f663d56-37e2-4a3f-9bb1-24e31b011470-host\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.834302 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.834566 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/6f663d56-37e2-4a3f-9bb1-24e31b011470-serviceca\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.853801 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czrsx\" (UniqueName: \"kubernetes.io/projected/6f663d56-37e2-4a3f-9bb1-24e31b011470-kube-api-access-czrsx\") pod \"node-ca-n6rc2\" (UID: \"6f663d56-37e2-4a3f-9bb1-24e31b011470\") " pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.856943 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.873232 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.888945 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.890991 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-n6rc2" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.901920 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.904223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.904269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.904287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.904309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.904326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4894]: W1210 18:55:52.912187 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f663d56_37e2_4a3f_9bb1_24e31b011470.slice/crio-dfe5bd223c8bbe02fb3d087a8f6abe79f6ad1f20c74d90096f3a9c51b3fb9d21 WatchSource:0}: Error finding container dfe5bd223c8bbe02fb3d087a8f6abe79f6ad1f20c74d90096f3a9c51b3fb9d21: Status 404 returned error can't find the container with id dfe5bd223c8bbe02fb3d087a8f6abe79f6ad1f20c74d90096f3a9c51b3fb9d21 Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.920727 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.936704 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.957484 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4894]: I1210 18:55:52.984360 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.011778 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.013499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.013537 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.013549 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.013565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.013575 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.024559 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.040978 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.055910 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.065154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.115346 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.115433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.115450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.115467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.115479 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.219312 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.219388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.219406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.219437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.219455 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.322451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.322581 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.322604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.322633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.322653 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.424961 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.425023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.425040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.425065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.425082 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.527653 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.527710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.527727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.527752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.527772 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.540064 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:53 crc kubenswrapper[4894]: E1210 18:55:53.540219 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.541264 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:53 crc kubenswrapper[4894]: E1210 18:55:53.541502 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.630311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.630359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.630372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.630389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.630403 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.733160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.733222 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.733241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.733265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.733281 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.757395 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-n6rc2" event={"ID":"6f663d56-37e2-4a3f-9bb1-24e31b011470","Type":"ContainerStarted","Data":"2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.757472 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-n6rc2" event={"ID":"6f663d56-37e2-4a3f-9bb1-24e31b011470","Type":"ContainerStarted","Data":"dfe5bd223c8bbe02fb3d087a8f6abe79f6ad1f20c74d90096f3a9c51b3fb9d21"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.763437 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.765824 4894 generic.go:334] "Generic (PLEG): container finished" podID="bb325131-20c4-4791-9bbc-45ae36443184" containerID="bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77" exitCode=0 Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.765892 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerDied","Data":"bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.778458 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.791216 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.807392 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.829153 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.835943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.835980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.835993 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.836011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.836021 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.843302 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.856963 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.870616 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.898579 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.912514 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.924832 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.938596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.938673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.938699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.938729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.938751 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.939830 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.950571 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.964757 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.976031 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4894]: I1210 18:55:53.988695 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.000515 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.015214 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.026584 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.041159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.041188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.041197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.041209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.041191 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.041218 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.053876 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.068651 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.077971 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.085833 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.111029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.122781 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.136286 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.143290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.143337 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.143346 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.143359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.143367 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.144320 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.160325 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.256097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.256140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.256153 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.256168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.256179 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.357968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.358004 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.358015 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.358031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.358043 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.460709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.460747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.460758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.460772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.460783 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.540822 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:54 crc kubenswrapper[4894]: E1210 18:55:54.540987 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.563531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.563591 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.563608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.563633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.563650 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.666530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.666768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.666899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.666986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.667078 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.770080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.770877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.770904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.770923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.770933 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.772018 4894 generic.go:334] "Generic (PLEG): container finished" podID="bb325131-20c4-4791-9bbc-45ae36443184" containerID="be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378" exitCode=0 Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.772070 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerDied","Data":"be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.784624 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.799970 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.810369 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.823715 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.842555 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.864347 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.873727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.873761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.873772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.873786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.873795 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.877615 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.903928 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.923160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.937505 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.948645 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.960891 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.974787 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.977343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.977382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.977398 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.977415 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.977439 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4894]: I1210 18:55:54.989624 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:54Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.080838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.080927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.080945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.080974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.080993 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.184293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.184340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.184357 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.184414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.184433 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.287193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.287254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.287275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.287304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.287326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.390121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.390183 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.390205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.390237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.390260 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.492774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.492829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.492877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.492901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.492919 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.540461 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.540462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:55 crc kubenswrapper[4894]: E1210 18:55:55.540638 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:55 crc kubenswrapper[4894]: E1210 18:55:55.540800 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.596285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.596340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.596357 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.596381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.596399 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.698931 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.699006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.699029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.699064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.699108 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.781594 4894 generic.go:334] "Generic (PLEG): container finished" podID="bb325131-20c4-4791-9bbc-45ae36443184" containerID="59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf" exitCode=0 Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.781713 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerDied","Data":"59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.802821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.802910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.802929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.802954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.802972 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.803308 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.820341 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.851827 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.873889 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.895288 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.906173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.906237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.906261 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.906293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.906317 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.910530 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.930812 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.949233 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.970105 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4894]: I1210 18:55:55.988476 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.009010 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.009551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.009583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.009593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.009609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.009621 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.024315 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.037456 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.048530 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.111573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.111633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.111651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.111675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.111695 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.215359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.215419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.215436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.215460 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.215479 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.318147 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.318180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.318191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.318205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.318216 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.420927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.420962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.420970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.420985 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.420993 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.523669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.523723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.523742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.523765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.523782 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.540391 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:56 crc kubenswrapper[4894]: E1210 18:55:56.540570 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.627475 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.627556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.627579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.627611 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.627637 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.730925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.730983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.731000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.731022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.731041 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.792638 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.793195 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.799181 4894 generic.go:334] "Generic (PLEG): container finished" podID="bb325131-20c4-4791-9bbc-45ae36443184" containerID="d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb" exitCode=0 Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.799247 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerDied","Data":"d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.809608 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.813945 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.830165 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.834209 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.837149 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.837208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.837242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.837269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.837289 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.861154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.881532 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.900616 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.916023 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.929266 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.940283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.940321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.940329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.940369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.940381 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.942909 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.961123 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.975887 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:56 crc kubenswrapper[4894]: I1210 18:55:56.989536 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:56Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.008491 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.029395 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.043115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.043148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.043161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.043180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.043191 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.044416 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.063279 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.075276 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.088217 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.103480 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.118453 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.131438 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.144551 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.146056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.146148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.146243 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.146307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.146370 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.169928 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.187390 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.205100 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.219279 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.234362 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.249633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.249681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.249703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.249731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.249754 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.256232 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.276827 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.353239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.353307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.353324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.353349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.353366 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.455907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.455942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.455951 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.455965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.455976 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.541190 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.541382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:57 crc kubenswrapper[4894]: E1210 18:55:57.542033 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:57 crc kubenswrapper[4894]: E1210 18:55:57.542266 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.558338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.558384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.558401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.558425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.558442 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.565057 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.583213 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.597605 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.613705 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.632365 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.650016 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.661304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.661361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.661384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.661417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.661441 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.674474 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.689957 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.702534 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.727266 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.745401 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.764390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.764436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.764447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.764463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.764475 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.767563 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.785394 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.801170 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.803581 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.803659 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.833984 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.883401 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.885093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.885146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.885168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.885189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.885207 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.916369 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.929000 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.939723 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.953518 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.967422 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.983909 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.987066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.987110 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.987121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.987142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.987154 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.993929 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:57 crc kubenswrapper[4894]: E1210 18:55:57.994266 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:56:13.994237894 +0000 UTC m=+56.789085696 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:57 crc kubenswrapper[4894]: I1210 18:55:57.997624 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:57Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.021691 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.039470 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.055107 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.068174 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.077102 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.088928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.088964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.088973 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.088986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.088995 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.093424 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.094641 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.094676 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.094705 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.094727 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094756 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094821 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094835 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094836 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094862 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:14.094822177 +0000 UTC m=+56.889669969 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094871 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094895 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:14.094884809 +0000 UTC m=+56.889732601 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094917 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:14.09490895 +0000 UTC m=+56.889756862 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094967 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094981 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.094995 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.095035 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:14.095021673 +0000 UTC m=+56.889869535 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.191388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.191627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.191692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.191754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.191816 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.294354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.294393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.294402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.294417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.294426 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.397697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.397781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.397799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.397875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.397893 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.501241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.501285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.501302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.501323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.501338 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.539996 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.540128 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.604249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.604296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.604308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.604327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.604339 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.706745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.706787 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.706798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.706814 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.706827 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.808922 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.808988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.809006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.809031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.809053 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.810334 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/0.log" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.810547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.810583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.810594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.810611 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.810625 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.814444 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843" exitCode=1 Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.814591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.815533 4894 scope.go:117] "RemoveContainer" containerID="d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.821204 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" event={"ID":"bb325131-20c4-4791-9bbc-45ae36443184","Type":"ContainerStarted","Data":"1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957"} Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.832536 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.837784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.837881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.837907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.837932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.837949 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.837924 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.853011 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.862452 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.865989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.866024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.866036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.866052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.866064 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.882364 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"space (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:58.648033 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 18:55:58.648079 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 18:55:58.648129 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:58.648128 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 18:55:58.648166 6228 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:58.648177 6228 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:58.648175 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 18:55:58.648216 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:55:58.648238 6228 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:58.648252 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 18:55:58.648254 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:55:58.648272 6228 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:58.648278 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:58.648279 6228 factory.go:656] Stopping watch factory\\\\nI1210 18:55:58.648300 6228 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.883980 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.889329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.889455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.889524 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.889667 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.889992 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.905789 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.906870 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.911222 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.911245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.911254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.911266 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.911274 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.923426 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.927436 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: E1210 18:55:58.927588 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.929172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.929209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.929220 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.929237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.929249 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.936733 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.951746 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.967187 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.979744 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:58 crc kubenswrapper[4894]: I1210 18:55:58.996576 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:58Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.024822 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.033313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.036984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.037151 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.037344 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.037597 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.041816 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.062807 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.075107 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.087363 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.098674 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.109385 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.122594 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.135948 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.139616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.139666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.139676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.139691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.139699 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.148658 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.172143 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"space (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:58.648033 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 18:55:58.648079 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 18:55:58.648129 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:58.648128 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 18:55:58.648166 6228 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:58.648177 6228 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:58.648175 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 18:55:58.648216 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:55:58.648238 6228 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:58.648252 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 18:55:58.648254 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:55:58.648272 6228 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:58.648278 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:58.648279 6228 factory.go:656] Stopping watch factory\\\\nI1210 18:55:58.648300 6228 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.188058 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.201318 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.214766 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.231388 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.241647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.241680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.241689 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.241702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.241712 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.249482 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.261424 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.272687 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.343879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.343923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.343937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.343955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.343967 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.446339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.446372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.446384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.446404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.446418 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.540483 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.540542 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:59 crc kubenswrapper[4894]: E1210 18:55:59.540623 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:59 crc kubenswrapper[4894]: E1210 18:55:59.540749 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.548127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.548156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.548165 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.548176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.548186 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.651102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.651133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.651145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.651161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.651170 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.754769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.754843 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.754884 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.754910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.754937 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.829465 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/0.log" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.833975 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.834662 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.855229 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.857403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.857490 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.857535 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.857553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.857563 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.872099 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.886305 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.906633 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.927072 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.947618 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.959384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.959429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.959442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.959459 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.959473 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.973112 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.986255 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4894]: I1210 18:55:59.997044 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.015339 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"space (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:58.648033 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 18:55:58.648079 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 18:55:58.648129 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:58.648128 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 18:55:58.648166 6228 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:58.648177 6228 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:58.648175 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 18:55:58.648216 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:55:58.648238 6228 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:58.648252 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 18:55:58.648254 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:55:58.648272 6228 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:58.648278 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:58.648279 6228 factory.go:656] Stopping watch factory\\\\nI1210 18:55:58.648300 6228 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.028331 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.042006 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.052795 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.061800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.061870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.061883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.061902 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.061914 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.069355 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.164336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.164374 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.164383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.164396 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.164405 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.268455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.268516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.268533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.268558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.268575 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.371093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.371176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.371202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.371236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.371261 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.473452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.473496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.473507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.473525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.473538 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.541259 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:00 crc kubenswrapper[4894]: E1210 18:56:00.541517 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.576258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.576318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.576339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.576366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.576384 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.679404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.679465 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.679483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.679518 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.679536 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.782945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.782994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.783009 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.783029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.783047 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.839621 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/1.log" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.841265 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/0.log" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.845205 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2" exitCode=1 Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.845250 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.845314 4894 scope.go:117] "RemoveContainer" containerID="d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.847270 4894 scope.go:117] "RemoveContainer" containerID="90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2" Dec 10 18:56:00 crc kubenswrapper[4894]: E1210 18:56:00.847817 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.868508 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.886258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.886339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.886366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.886396 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.886424 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.890229 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.903404 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.923175 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.937573 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.955152 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.971595 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.989074 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.989413 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.989587 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.989762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.989996 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4894]: I1210 18:56:00.992837 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.006628 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.022247 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.042542 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.061251 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.074075 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.092435 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.092472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.092485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.092505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.092517 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.096703 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"space (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:58.648033 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 18:55:58.648079 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 18:55:58.648129 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:58.648128 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 18:55:58.648166 6228 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:58.648177 6228 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:58.648175 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 18:55:58.648216 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:55:58.648238 6228 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:58.648252 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 18:55:58.648254 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:55:58.648272 6228 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:58.648278 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:58.648279 6228 factory.go:656] Stopping watch factory\\\\nI1210 18:55:58.648300 6228 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.194936 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.194990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.195007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.195032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.195051 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.283582 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp"] Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.284331 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.287266 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.288741 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.298759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.298831 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.298890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.298923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.299181 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.306958 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.325121 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.328948 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vwtx\" (UniqueName: \"kubernetes.io/projected/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-kube-api-access-2vwtx\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.329072 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.329117 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.329176 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.347238 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.361486 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.374650 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.388650 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.401334 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.401903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.401962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.401981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.402005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.402024 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.419793 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.430117 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vwtx\" (UniqueName: \"kubernetes.io/projected/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-kube-api-access-2vwtx\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.430287 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.430386 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.430486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.431479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.431900 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.443567 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.469419 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.470448 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vwtx\" (UniqueName: \"kubernetes.io/projected/8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce-kube-api-access-2vwtx\") pod \"ovnkube-control-plane-749d76644c-bxsrp\" (UID: \"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.487040 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.501546 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.503904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.503972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.503995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.504026 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.504049 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.523568 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d71117ef20cc30e0252147f9b6930241c8925d49ac8ab908d0c74b90ddf3f843\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"message\\\":\\\"space (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:58.648033 6228 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 18:55:58.648079 6228 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 18:55:58.648129 6228 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:58.648128 6228 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 18:55:58.648166 6228 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:58.648177 6228 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:58.648175 6228 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 18:55:58.648216 6228 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:55:58.648238 6228 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:58.648252 6228 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 18:55:58.648254 6228 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:55:58.648272 6228 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:58.648278 6228 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:58.648279 6228 factory.go:656] Stopping watch factory\\\\nI1210 18:55:58.648300 6228 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.536167 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.541013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:01 crc kubenswrapper[4894]: E1210 18:56:01.541232 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.541699 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:01 crc kubenswrapper[4894]: E1210 18:56:01.541878 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.552688 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.566428 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.603722 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.606642 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.606703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.606720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.606745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.606766 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: W1210 18:56:01.622133 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d85c5b7_37ba_4aaa_a1d9_52882a94e0ce.slice/crio-14d44f9011aa07e06dc14218a4794900ad642018e572e83a0b5afdcae386d00b WatchSource:0}: Error finding container 14d44f9011aa07e06dc14218a4794900ad642018e572e83a0b5afdcae386d00b: Status 404 returned error can't find the container with id 14d44f9011aa07e06dc14218a4794900ad642018e572e83a0b5afdcae386d00b Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.711131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.711556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.712056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.712250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.712422 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.816447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.816502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.816515 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.816544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.816558 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.851138 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/1.log" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.854390 4894 scope.go:117] "RemoveContainer" containerID="90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2" Dec 10 18:56:01 crc kubenswrapper[4894]: E1210 18:56:01.854562 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.855184 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" event={"ID":"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce","Type":"ContainerStarted","Data":"14d44f9011aa07e06dc14218a4794900ad642018e572e83a0b5afdcae386d00b"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.868767 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.879467 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.896514 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.910830 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.922485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.922527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.922538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.922556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.922568 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.926446 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.937291 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.950642 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.969235 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4894]: I1210 18:56:01.987531 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.002900 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.017872 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.024987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.025043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.025060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.025078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.025087 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.031285 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.045737 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.058672 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.079730 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.127997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.128030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.128039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.128052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.128060 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.230404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.230454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.230467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.230483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.230493 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.332696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.332779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.332790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.332806 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.332816 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.434571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.434634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.434651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.434676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.434697 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.537472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.537530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.537553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.537585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.537608 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.540760 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:02 crc kubenswrapper[4894]: E1210 18:56:02.540971 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.640521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.640781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.640885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.640988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.641127 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.747422 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.747467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.747482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.747502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.747517 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.850018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.850073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.850087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.850106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.850119 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.859345 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" event={"ID":"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce","Type":"ContainerStarted","Data":"49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.859388 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" event={"ID":"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce","Type":"ContainerStarted","Data":"b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.873911 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.888167 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.902899 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.919624 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.937414 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.952190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.952239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.952269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.952295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.952312 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.953906 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.972687 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4894]: I1210 18:56:02.984997 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.007518 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.026634 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.039370 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.054391 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.054451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.054468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.054493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.054512 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.062327 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.083575 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.103027 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.116664 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.128990 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7dks9"] Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.129840 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.129999 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.147910 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.152715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.152772 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8szbg\" (UniqueName: \"kubernetes.io/projected/72521f10-d803-4b10-97dc-cbad9bda6be7-kube-api-access-8szbg\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.157843 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.157928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.157944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.157969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.157985 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.162280 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.201170 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.232527 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.253723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.253768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8szbg\" (UniqueName: \"kubernetes.io/projected/72521f10-d803-4b10-97dc-cbad9bda6be7-kube-api-access-8szbg\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.253918 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.253999 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:03.753978151 +0000 UTC m=+46.548825953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.260136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.260185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.260200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.260220 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.260235 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.267332 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.276017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8szbg\" (UniqueName: \"kubernetes.io/projected/72521f10-d803-4b10-97dc-cbad9bda6be7-kube-api-access-8szbg\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.285936 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.305197 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.316570 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.331323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.340709 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.349947 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.361747 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.362604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.362654 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.362672 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.362697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.362715 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.373783 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.392110 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.415460 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.429816 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:03Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.464891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.465081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.465109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.465137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.465154 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.540987 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.540987 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.541227 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.541352 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.568281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.568356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.568387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.568418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.568440 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.671640 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.671686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.671703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.671725 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.671742 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.757584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.757798 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:03 crc kubenswrapper[4894]: E1210 18:56:03.757926 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:04.757900866 +0000 UTC m=+47.552748688 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.774292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.774350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.774376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.774407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.774430 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.877033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.877088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.877106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.877131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.877149 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.980050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.980137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.980159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.980182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4894]: I1210 18:56:03.980261 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.083484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.083601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.083640 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.083671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.083693 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.186983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.187051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.187104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.187134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.187151 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.290339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.290754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.290964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.291126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.291279 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.395234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.395296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.395319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.395346 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.395366 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.498325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.498373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.498385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.498403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.498416 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.540983 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.541011 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:04 crc kubenswrapper[4894]: E1210 18:56:04.541194 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:04 crc kubenswrapper[4894]: E1210 18:56:04.541453 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.602478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.602527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.602548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.602576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.602598 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.706258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.706566 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.706737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.706940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.707162 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.770507 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:04 crc kubenswrapper[4894]: E1210 18:56:04.770720 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:04 crc kubenswrapper[4894]: E1210 18:56:04.770800 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:06.770777892 +0000 UTC m=+49.565625714 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.810640 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.810684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.810697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.810712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.810724 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.913403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.913467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.913488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.913516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4894]: I1210 18:56:04.913534 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.016514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.016570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.016587 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.016611 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.016628 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.119593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.119641 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.119658 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.119688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.119710 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.225416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.225468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.225489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.225516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.225536 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.328057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.328078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.328086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.328146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.328155 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.431059 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.431088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.431099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.431116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.431126 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.533898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.533967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.533985 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.534016 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.534034 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.540448 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.540502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:05 crc kubenswrapper[4894]: E1210 18:56:05.540622 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:05 crc kubenswrapper[4894]: E1210 18:56:05.540796 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.636958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.637017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.637040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.637072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.637095 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.739899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.739976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.740003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.740033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.740054 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.843682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.843747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.843768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.843797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.843818 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.946586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.946636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.946652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.946674 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4894]: I1210 18:56:05.946690 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.050227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.050267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.050284 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.050307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.050325 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.153063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.153131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.153154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.153182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.153202 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.256343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.256418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.256442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.256470 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.256492 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.358984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.359048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.359077 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.359104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.359125 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.461469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.461518 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.461530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.461545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.461556 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.540605 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.540656 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:06 crc kubenswrapper[4894]: E1210 18:56:06.540804 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:06 crc kubenswrapper[4894]: E1210 18:56:06.540981 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.565577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.565632 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.565681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.565702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.565721 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.669634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.669691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.669708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.669730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.669746 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.773238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.773358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.773420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.773451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.773472 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.793046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:06 crc kubenswrapper[4894]: E1210 18:56:06.793243 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:06 crc kubenswrapper[4894]: E1210 18:56:06.793364 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:10.793332415 +0000 UTC m=+53.588180247 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.877094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.877184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.877201 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.877268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.877293 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.980242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.980329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.980348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.980375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4894]: I1210 18:56:06.980393 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.084057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.084122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.084140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.084174 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.084193 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.187636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.187710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.187730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.187758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.187779 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.291541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.291587 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.291605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.291629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.291647 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.394397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.394445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.394464 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.394509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.394536 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.497991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.498090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.498124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.498152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.498171 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.511077 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.524122 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.534497 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.540405 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.540616 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:07 crc kubenswrapper[4894]: E1210 18:56:07.540740 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:07 crc kubenswrapper[4894]: E1210 18:56:07.541123 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.550586 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.582165 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.599742 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.601293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.601335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.601353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.601379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.601396 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.615426 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.630886 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.652093 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.669617 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.681276 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.697794 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.704616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.704814 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.704916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.705023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.705112 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.718240 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.737009 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.756605 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.782471 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.798185 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.807227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.807268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.807282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.807302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.807318 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.815367 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.832452 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.844472 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.869742 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.887925 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.903072 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.910383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.910423 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.910437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.910456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.910472 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.915166 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.933994 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.948685 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.961521 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.975615 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4894]: I1210 18:56:07.993673 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.007934 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.012958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.012981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.012990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.013009 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.013019 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.025983 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.046770 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.063737 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.074503 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.085993 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.116301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.116352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.116362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.116379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.116394 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.219989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.220062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.220090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.220123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.220149 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.323539 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.323615 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.323633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.323656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.323673 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.426449 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.426526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.426548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.426578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.426602 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.530320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.530390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.530458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.530500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.530525 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.540916 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.540965 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:08 crc kubenswrapper[4894]: E1210 18:56:08.541085 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:08 crc kubenswrapper[4894]: E1210 18:56:08.541193 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.633676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.633743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.633760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.633785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.633803 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.736952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.737038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.737048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.737067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.737079 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.840902 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.840955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.840978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.841028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.841054 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.944421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.944500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.944526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.944554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.944576 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.980761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.980820 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.980888 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.980925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4894]: I1210 18:56:08.980946 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.001913 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.007216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.007297 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.007317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.007353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.007377 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.026977 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.030756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.030808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.030821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.030840 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.030871 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.042330 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.047296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.047368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.047387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.047412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.047434 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.068220 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.073818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.073946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.073971 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.074008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.074033 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.095007 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.095149 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.096953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.097019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.097042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.097069 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.097088 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.200215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.200283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.200305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.200331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.200350 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.303800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.303891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.303910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.303935 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.303953 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.406771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.406828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.406869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.406895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.406914 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.510337 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.511228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.511447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.511628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.511819 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.540202 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.540209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.540336 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:09 crc kubenswrapper[4894]: E1210 18:56:09.540398 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.617003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.617072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.617102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.617128 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.617145 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.719908 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.719950 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.719967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.719990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.720006 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.822940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.823018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.823037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.823062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.823079 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.926218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.926287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.926311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.926340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4894]: I1210 18:56:09.926362 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.029101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.029148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.029165 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.029190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.029208 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.132182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.132239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.132255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.132281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.132304 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.235577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.235642 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.235660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.235686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.235705 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.339022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.339083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.339100 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.339124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.339142 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.442215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.442287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.442311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.442339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.442361 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.540451 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.540451 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:10 crc kubenswrapper[4894]: E1210 18:56:10.540641 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:10 crc kubenswrapper[4894]: E1210 18:56:10.540748 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.545012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.545082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.545106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.545135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.545160 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.648360 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.648416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.648434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.648457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.648476 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.751021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.751074 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.751088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.751108 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.751120 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.834101 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:10 crc kubenswrapper[4894]: E1210 18:56:10.834283 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:10 crc kubenswrapper[4894]: E1210 18:56:10.834405 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:18.834373085 +0000 UTC m=+61.629220917 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.853799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.853891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.853930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.853963 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.853987 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.957258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.957349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.957373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.957399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4894]: I1210 18:56:10.957421 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.060982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.061054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.061072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.061102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.061123 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.164746 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.164826 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.164883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.164919 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.164943 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.268274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.268362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.268381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.268408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.268428 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.371410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.371470 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.371486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.371511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.371533 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.474371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.474427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.474445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.474468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.474485 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.540469 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:11 crc kubenswrapper[4894]: E1210 18:56:11.540646 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.541194 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:11 crc kubenswrapper[4894]: E1210 18:56:11.541416 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.577495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.577550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.577568 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.577593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.577610 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.680509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.680545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.680556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.680571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.680583 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.783457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.783509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.783526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.783549 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.783568 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.886406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.886468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.886494 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.886522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.886543 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.989174 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.989231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.989254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.989281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4894]: I1210 18:56:11.989302 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.092408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.092477 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.092495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.092527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.092546 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.195596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.199058 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.199081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.199131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.199156 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.302938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.302997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.303014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.303036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.303051 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.406429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.406494 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.406513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.406538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.406555 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.509745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.509814 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.509835 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.509898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.509916 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.540476 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.540476 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:12 crc kubenswrapper[4894]: E1210 18:56:12.540675 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:12 crc kubenswrapper[4894]: E1210 18:56:12.540896 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.612982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.613043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.613062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.613088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.613108 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.716533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.716592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.716609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.716632 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.716650 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.819651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.819698 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.819715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.819738 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.819755 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.922678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.922774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.922801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.922836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4894]: I1210 18:56:12.922915 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.026209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.026271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.026291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.026313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.026330 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.129432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.129487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.129504 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.129526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.129544 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.233338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.233389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.233405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.233428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.233444 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.336314 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.336383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.336403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.336427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.336444 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.439628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.439676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.439693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.439717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.439732 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.540944 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.541030 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:13 crc kubenswrapper[4894]: E1210 18:56:13.541166 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:13 crc kubenswrapper[4894]: E1210 18:56:13.541519 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.542636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.542672 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.542681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.542696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.542707 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.646407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.646493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.646517 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.646550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.646575 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.749614 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.749673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.749810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.749887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.749931 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.852798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.852904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.852928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.852960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.852981 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.955883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.955961 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.955985 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.956017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4894]: I1210 18:56:13.956039 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.058932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.059009 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.059042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.059105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.059126 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.072515 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.072713 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:56:46.072674975 +0000 UTC m=+88.867522807 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.162482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.162541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.162558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.162581 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.162599 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.174533 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.174596 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.174643 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.174717 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174884 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174914 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174938 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174943 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174957 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174968 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.174882 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.175025 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.175021 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:46.174999332 +0000 UTC m=+88.969847164 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.175129 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:46.175105255 +0000 UTC m=+88.969953167 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.175151 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:46.175141796 +0000 UTC m=+88.969989728 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.175167 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:46.175158706 +0000 UTC m=+88.970006648 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.265290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.265351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.265368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.265391 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.265408 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.368942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.369008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.369029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.369058 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.369080 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.471723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.471776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.471801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.471822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.471838 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.540832 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.540936 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.541027 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:14 crc kubenswrapper[4894]: E1210 18:56:14.541077 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.541747 4894 scope.go:117] "RemoveContainer" containerID="90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.575716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.575753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.575767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.575787 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.575801 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.679231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.680428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.680808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.681196 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.681648 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.784769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.784811 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.784827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.784873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.784890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.887765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.887835 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.887890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.887934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.887958 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.902584 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/1.log" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.906477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.907627 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.944833 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:14Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.969323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:14Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.990600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.990662 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.990687 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.990716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.990737 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4894]: I1210 18:56:14.993272 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:14Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.009411 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.038934 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.063081 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.081417 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.093132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.093166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.093177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.093194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.093205 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.097771 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.108588 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.118778 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.129433 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.140341 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.152867 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.166266 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.183918 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.195191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.195236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.195249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.195267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.195281 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.198012 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.212408 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.297554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.297610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.297629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.297684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.297703 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.400468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.400539 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.400558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.400577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.400620 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.503003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.503030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.503037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.503051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.503062 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.541012 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:15 crc kubenswrapper[4894]: E1210 18:56:15.541151 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.541585 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:15 crc kubenswrapper[4894]: E1210 18:56:15.541695 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.605696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.605757 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.605768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.605783 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.605794 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.709217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.709368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.709396 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.709471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.709596 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.813365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.813446 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.813471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.813555 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.814262 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.913529 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/2.log" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.914827 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/1.log" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.917106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.917153 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.917170 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.917194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.917211 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.919906 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6" exitCode=1 Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.919959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6"} Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.920013 4894 scope.go:117] "RemoveContainer" containerID="90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.921688 4894 scope.go:117] "RemoveContainer" containerID="f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6" Dec 10 18:56:15 crc kubenswrapper[4894]: E1210 18:56:15.922302 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.946591 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.960761 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:15 crc kubenswrapper[4894]: I1210 18:56:15.988640 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90bbd7916358138b007941088aec7d7e6c30313de32afd1ab4aa8481b1de06f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:59Z\\\",\\\"message\\\":\\\"ices.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 18:55:59.601547 6378 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/catalog-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/catalog-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.204:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {78f6184b-c7cf-436d-8cbb-4b31f8af75e8}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:59.601821 63\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:15Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.007794 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.020707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.020906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.020929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.020953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.020970 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.027227 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.041214 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.060628 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.077447 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.091443 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.105811 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.123678 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.124988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.125064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.125088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.125119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.125142 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.143092 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.161791 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.180554 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.196601 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.211892 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.228418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.228484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.228509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.228540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.228564 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.232201 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.331891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.331942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.331953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.331972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.331984 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.434911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.434974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.434992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.435018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.435037 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.538197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.538251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.538264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.538285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.538298 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.540547 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:16 crc kubenswrapper[4894]: E1210 18:56:16.540682 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.540949 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:16 crc kubenswrapper[4894]: E1210 18:56:16.541056 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.641212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.641256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.641266 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.641281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.641293 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.744309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.744392 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.744409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.744434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.744458 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.847423 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.847509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.847532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.847561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.847589 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.927516 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/2.log" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.933712 4894 scope.go:117] "RemoveContainer" containerID="f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6" Dec 10 18:56:16 crc kubenswrapper[4894]: E1210 18:56:16.934094 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.951062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.951137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.951160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.951190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.951217 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.959314 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:16 crc kubenswrapper[4894]: I1210 18:56:16.979716 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:16Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.011287 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.034160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.050018 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.057384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.057418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.057429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.057444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.057456 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.063264 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.077272 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.091482 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.103987 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.120518 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.136077 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.148908 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.160531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.160596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.160616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.161524 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.161549 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.163052 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.180502 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.194833 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.219760 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.237174 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.264371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.264439 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.264457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.264484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.264503 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.368233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.368318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.368342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.368369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.368387 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.471679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.471765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.471784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.471809 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.471827 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.539976 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:17 crc kubenswrapper[4894]: E1210 18:56:17.540181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.543909 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:17 crc kubenswrapper[4894]: E1210 18:56:17.544259 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.562081 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.577192 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.577772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.577794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.577901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.577921 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.585170 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.610801 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.628653 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.647607 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.667467 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.680901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.680965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.680982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.681007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.681027 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.684816 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.699428 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.717619 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.739282 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.758116 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.782985 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.783057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.783069 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.783087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.783098 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.787615 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.807150 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.820012 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.833487 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.854215 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.872372 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:17Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.885469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.885534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.885553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.885578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.885597 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.989036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.989112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.989133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.989164 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4894]: I1210 18:56:17.989187 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.092322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.092394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.092411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.092439 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.092457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.196444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.196492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.196511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.196534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.196550 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.299932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.299999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.300025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.300056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.300081 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.403157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.403216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.403237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.403259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.403277 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.505383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.505451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.505482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.505512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.505533 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.540307 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.540372 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:18 crc kubenswrapper[4894]: E1210 18:56:18.540501 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:18 crc kubenswrapper[4894]: E1210 18:56:18.540685 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.608873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.608937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.608956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.608979 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.608996 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.711258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.711319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.711336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.711359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.711375 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.815656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.815731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.815754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.815781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.815805 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.919822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.919890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.919903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.919920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.919933 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4894]: I1210 18:56:18.927577 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:18 crc kubenswrapper[4894]: E1210 18:56:18.927699 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:18 crc kubenswrapper[4894]: E1210 18:56:18.927751 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:34.927735036 +0000 UTC m=+77.722582838 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.023090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.023153 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.023169 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.023193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.023214 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.126180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.126254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.126271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.126295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.126312 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.228927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.228986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.229005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.229030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.229048 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.331377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.331411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.331418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.331432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.331442 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.419656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.419748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.419786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.419816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.419840 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.444090 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.448558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.448613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.448623 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.448636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.448645 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.463823 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.468521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.468588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.468599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.468614 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.468623 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.484573 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.488945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.488985 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.489000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.489024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.489041 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.507040 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.512305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.512398 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.512418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.512443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.512461 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.529810 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.530081 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.531641 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.531677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.531686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.531700 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.531710 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.540180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.540302 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.540408 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:19 crc kubenswrapper[4894]: E1210 18:56:19.540566 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.633834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.633945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.633962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.633990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.634007 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.736723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.736772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.736809 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.736828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.736904 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.844590 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.844640 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.844656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.844680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.844697 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.946973 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.947045 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.947063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.947085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4894]: I1210 18:56:19.947101 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.091916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.091986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.092008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.092040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.092063 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.195232 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.195299 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.195316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.195339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.195357 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.298634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.298688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.298703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.298730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.298747 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.408365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.408418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.408434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.408458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.408475 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.511444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.511510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.511529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.511554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.511574 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.540151 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.540211 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:20 crc kubenswrapper[4894]: E1210 18:56:20.540327 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:20 crc kubenswrapper[4894]: E1210 18:56:20.540465 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.614393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.614569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.614593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.614621 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.614643 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.717632 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.717695 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.717708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.717729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.717747 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.820810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.820858 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.820866 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.820879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.820887 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.924194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.924259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.924324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.924352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4894]: I1210 18:56:20.924370 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.028207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.028496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.028512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.028536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.028552 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.131322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.131389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.131414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.131443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.131465 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.233903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.233964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.233977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.233992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.234002 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.336446 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.336499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.336519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.336541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.336557 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.438378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.438403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.438411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.438422 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.438430 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.539924 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.540013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:21 crc kubenswrapper[4894]: E1210 18:56:21.540103 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:21 crc kubenswrapper[4894]: E1210 18:56:21.540262 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.541086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.541112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.541120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.541131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.541142 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.644109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.644148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.644160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.644177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.644187 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.748299 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.748359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.748375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.748399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.748416 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.851488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.851545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.851561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.851584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.851600 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.953786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.953875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.953899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.953926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4894]: I1210 18:56:21.953949 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.056228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.056293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.056311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.056336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.056357 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.158948 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.158983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.158995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.159011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.159023 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.261353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.261404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.261420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.261441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.261457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.364152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.364193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.364204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.364219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.364230 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.466526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.466575 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.466584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.466599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.466609 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.540327 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:22 crc kubenswrapper[4894]: E1210 18:56:22.540475 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.540327 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:22 crc kubenswrapper[4894]: E1210 18:56:22.540554 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.568833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.568934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.568956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.568988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.569036 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.670739 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.670785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.670801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.670822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.670837 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.773514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.773563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.773579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.773603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.773619 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.875717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.875759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.875768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.875782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.875792 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.978280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.978320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.978330 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.978345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4894]: I1210 18:56:22.978355 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.081182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.081238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.081254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.081276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.081291 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.183757 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.183798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.183809 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.183823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.183833 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.285927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.285970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.285997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.286014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.286028 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.388862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.388928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.388938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.388955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.388967 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.491344 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.491383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.491393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.491406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.491415 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.540765 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.540812 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:23 crc kubenswrapper[4894]: E1210 18:56:23.540898 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:23 crc kubenswrapper[4894]: E1210 18:56:23.541076 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.593968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.594036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.594048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.594069 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.594083 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.695923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.695959 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.695970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.695984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.695994 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.799001 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.799044 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.799054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.799072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.799084 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.901916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.901953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.901965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.901981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4894]: I1210 18:56:23.901995 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.004762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.004807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.004816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.004830 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.004840 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.107334 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.107382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.107394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.107410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.107422 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.210209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.210249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.210257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.210270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.210278 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.312542 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.312573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.312582 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.312594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.312605 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.414135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.414166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.414177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.414193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.414205 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.516934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.516996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.517017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.517046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.517067 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.540894 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.540915 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:24 crc kubenswrapper[4894]: E1210 18:56:24.541069 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:24 crc kubenswrapper[4894]: E1210 18:56:24.541213 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.619765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.619825 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.619842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.619893 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.619910 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.723295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.723356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.723368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.723384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.723397 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.826097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.826161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.826180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.826207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.826225 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.929119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.929193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.929214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.929243 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4894]: I1210 18:56:24.929264 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.031040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.031111 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.031127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.031152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.031172 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.133705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.133756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.133774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.133797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.133815 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.235933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.235993 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.236013 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.236041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.236061 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.338323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.338356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.338364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.338396 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.338406 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.440596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.440626 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.440651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.440664 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.440674 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.540655 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.540696 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:25 crc kubenswrapper[4894]: E1210 18:56:25.540768 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:25 crc kubenswrapper[4894]: E1210 18:56:25.540922 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.542281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.542331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.542339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.542355 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.542366 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.645483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.645540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.645557 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.645579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.645603 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.748719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.748911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.748940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.748969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.748990 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.852046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.852134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.852151 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.852175 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.852191 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.955191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.955253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.955276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.955304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4894]: I1210 18:56:25.955326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.058553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.058592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.058602 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.058633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.058645 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.161163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.161226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.161237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.161271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.161285 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.264477 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.264536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.264553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.264577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.264595 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.366966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.367029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.367046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.367070 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.367087 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.469974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.470019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.470028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.470042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.470078 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.540289 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.540327 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:26 crc kubenswrapper[4894]: E1210 18:56:26.540428 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:26 crc kubenswrapper[4894]: E1210 18:56:26.540681 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.572373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.572410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.572420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.572438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.572450 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.675940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.676006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.676028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.676082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.676127 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.779646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.779739 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.779751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.779770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.779786 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.883209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.883274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.883291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.883316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.883338 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.985637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.985731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.985748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.985776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4894]: I1210 18:56:26.985795 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.089114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.089190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.089208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.089234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.089251 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.191280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.191311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.191319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.191333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.191341 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.293421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.293460 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.293468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.293550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.293565 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.395916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.395964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.395972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.395988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.395997 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.498280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.498320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.498332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.498348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.498358 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.540108 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.540166 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:27 crc kubenswrapper[4894]: E1210 18:56:27.540242 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:27 crc kubenswrapper[4894]: E1210 18:56:27.540437 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.552968 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.562099 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.577400 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.589647 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.601005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.601037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.601047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.601061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.601083 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.601761 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.610500 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.622201 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.633935 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.643832 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.656194 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.692665 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.704002 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.704067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.704087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.704105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.704116 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.717187 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.739006 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.750403 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.762169 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.772917 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.785557 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.806112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.806137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.806147 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.806162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.806171 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.907753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.907792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.907802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.907817 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4894]: I1210 18:56:27.907829 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.010010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.010069 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.010086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.010109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.010126 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.112185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.112246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.112264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.112287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.112303 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.215711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.215758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.215769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.215785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.215799 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.318768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.318820 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.318834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.318873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.318889 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.422173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.422323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.422345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.422368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.422385 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.525257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.525316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.525338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.525389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.525416 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.540809 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:28 crc kubenswrapper[4894]: E1210 18:56:28.540965 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.540814 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:28 crc kubenswrapper[4894]: E1210 18:56:28.541194 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.627973 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.628012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.628021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.628057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.628068 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.730828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.730878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.730888 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.730936 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.730945 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.833026 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.833085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.833093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.833123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.833133 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.935880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.935943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.935965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.935997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4894]: I1210 18:56:28.936017 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.038142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.038170 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.038178 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.038191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.038200 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.140101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.140139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.140148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.140162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.140171 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.242701 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.242815 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.242837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.242929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.242950 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.346312 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.346382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.346406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.346438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.346461 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.448249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.448498 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.448570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.448630 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.448691 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.540357 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.540659 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.540771 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.540984 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.555894 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.555943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.555961 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.555979 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.556020 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.556079 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.658378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.658412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.658421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.658434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.658443 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.761965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.762006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.762017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.762034 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.762046 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.795610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.795665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.795675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.795688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.795706 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.814860 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.819348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.819382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.819392 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.819406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.819417 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.833170 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.836966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.837003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.837015 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.837030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.837039 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.852398 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.857404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.857442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.857457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.857497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.857507 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.873777 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.879197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.879444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.879572 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.879718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.879854 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.900521 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4894]: E1210 18:56:29.900637 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.902241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.902264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.902275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.902290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4894]: I1210 18:56:29.902301 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.004577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.004647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.004669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.004696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.004715 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.107172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.107255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.107280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.107315 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.107344 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.210202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.210269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.210291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.210322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.210347 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.312329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.312399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.312417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.312441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.312459 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.415969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.416039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.416056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.416082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.416099 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.518531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.518605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.518622 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.518646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.518664 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.540053 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.540062 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:30 crc kubenswrapper[4894]: E1210 18:56:30.540303 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:30 crc kubenswrapper[4894]: E1210 18:56:30.540327 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.621673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.621735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.621754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.621780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.621797 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.724752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.724825 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.724884 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.724914 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.724935 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.826752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.827117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.827226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.827317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.827406 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.934176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.934311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.934340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.934394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4894]: I1210 18:56:30.934422 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.037680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.037999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.038121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.038265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.038444 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.141006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.141381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.141529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.141663 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.141789 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.245424 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.245716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.245837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.246067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.246200 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.348661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.348734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.348759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.348789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.348810 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.451550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.451603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.451620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.451642 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.451658 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.539926 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.539965 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:31 crc kubenswrapper[4894]: E1210 18:56:31.540094 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:31 crc kubenswrapper[4894]: E1210 18:56:31.540394 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.541592 4894 scope.go:117] "RemoveContainer" containerID="f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6" Dec 10 18:56:31 crc kubenswrapper[4894]: E1210 18:56:31.542166 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.554145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.554190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.554205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.554226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.554241 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.658213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.658968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.659003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.659033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.659060 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.761543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.761567 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.761585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.761597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.761606 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.863898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.863933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.863941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.863956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.863965 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.967216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.967301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.967323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.967350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4894]: I1210 18:56:31.967373 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.070543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.070616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.070638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.070665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.070685 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.173683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.173734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.173751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.173773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.173791 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.276474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.276528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.276546 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.276569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.276588 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.379604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.379668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.379693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.379726 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.379749 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.482647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.482717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.482740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.482769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.482790 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.540186 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.540193 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:32 crc kubenswrapper[4894]: E1210 18:56:32.540344 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:32 crc kubenswrapper[4894]: E1210 18:56:32.540473 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.585256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.585308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.585325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.585349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.585366 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.687808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.687837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.687859 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.687872 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.687881 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.791194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.791282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.791362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.791395 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.791429 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.895290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.895361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.895378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.895404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.895421 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.999160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.999242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.999286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.999321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4894]: I1210 18:56:32.999349 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.102530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.102588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.102603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.102625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.102640 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.206248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.206425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.206445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.206468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.206486 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.309130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.309198 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.309218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.309245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.309266 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.412489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.412554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.412574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.412602 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.412627 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.516151 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.516487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.516669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.516895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.517080 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.541141 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.541220 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:33 crc kubenswrapper[4894]: E1210 18:56:33.541387 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:33 crc kubenswrapper[4894]: E1210 18:56:33.541518 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.620645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.620703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.620720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.620746 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.620763 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.723785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.724765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.724973 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.725211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.725408 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.852856 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.852945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.852965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.852995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.853017 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.956140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.956197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.956215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.956239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4894]: I1210 18:56:33.956260 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.059300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.059351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.059371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.059393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.059410 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.163146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.163228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.163251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.163280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.163302 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.265755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.265815 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.265837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.265926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.265949 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.369038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.369096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.369113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.369140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.369158 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.471464 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.471914 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.472085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.472272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.472409 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.540581 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.540590 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:34 crc kubenswrapper[4894]: E1210 18:56:34.540996 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:34 crc kubenswrapper[4894]: E1210 18:56:34.541283 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.575468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.575741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.575931 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.576079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.576244 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.679194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.679280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.679302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.679324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.679340 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.781461 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.781528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.781548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.781574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.781594 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.884388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.884440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.884451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.884466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.884476 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.987943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.988278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.988407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.988727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.988906 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4894]: I1210 18:56:34.993484 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:34 crc kubenswrapper[4894]: E1210 18:56:34.993731 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:34 crc kubenswrapper[4894]: E1210 18:56:34.993887 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:06.993841911 +0000 UTC m=+109.788689713 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.093184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.093248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.093267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.093311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.093342 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.196897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.197307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.197440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.197579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.197744 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.302046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.302126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.302148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.302176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.302206 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.405880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.405941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.405958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.405983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.406001 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.509326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.509382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.509402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.509426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.509444 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.540638 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.540724 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:35 crc kubenswrapper[4894]: E1210 18:56:35.540806 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:35 crc kubenswrapper[4894]: E1210 18:56:35.540929 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.612637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.613061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.613309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.613573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.613764 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.717388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.717508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.717532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.717558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.717578 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.820250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.821070 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.821103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.821132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.821149 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.924545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.924982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.925319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.925414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.925490 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.995658 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/0.log" Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.996002 4894 generic.go:334] "Generic (PLEG): container finished" podID="d9392e0e-ca87-437a-8ba1-9c5d34a50660" containerID="a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d" exitCode=1 Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.996060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerDied","Data":"a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d"} Dec 10 18:56:35 crc kubenswrapper[4894]: I1210 18:56:35.996709 4894 scope.go:117] "RemoveContainer" containerID="a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.020335 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.029596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.029660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.029683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.029771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.029838 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.034122 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.058687 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.080293 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.099515 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.120174 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.133789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.134076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.134286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.134428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.134637 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.140805 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.156642 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.177773 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.191296 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.206947 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.220200 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.238079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.238137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.238160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.238190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.238208 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.239574 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.250232 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.264378 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.273916 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.292853 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.308320 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:36Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.340946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.340987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.341002 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.341025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.341042 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.443163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.443383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.443596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.443668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.443735 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.540637 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:36 crc kubenswrapper[4894]: E1210 18:56:36.540841 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.540644 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:36 crc kubenswrapper[4894]: E1210 18:56:36.541012 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.548160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.548405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.548593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.548765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.548945 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.652380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.652693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.652821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.653028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.653158 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.756508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.757407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.757583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.757744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.757974 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.861946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.863578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.864061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.864525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.864977 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.968101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.968473 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.968699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.968836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4894]: I1210 18:56:36.969061 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.004723 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/0.log" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.005202 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerStarted","Data":"dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.024248 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.045822 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.067634 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.073165 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.073214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.073229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.073250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.073270 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.089157 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.114342 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.133452 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.153394 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.169316 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.175583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.175649 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.175659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.175673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.175682 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.203831 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.226237 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.247885 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.263995 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.278416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.278467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.278487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.278511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.278528 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.282334 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.301474 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.323032 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.344688 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.363919 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.381485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.381556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.381576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.381601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.381618 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.382767 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.484930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.484970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.484980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.484996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.485008 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.540528 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:37 crc kubenswrapper[4894]: E1210 18:56:37.541026 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.540746 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:37 crc kubenswrapper[4894]: E1210 18:56:37.542363 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.567471 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.588749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.588801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.588819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.588881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.588899 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.590722 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.608449 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.624052 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.642181 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.658766 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.678174 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.692387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.692469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.692492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.692525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.692549 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.698018 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.714515 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.739781 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.758689 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.773179 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.793134 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.794491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.794536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.794548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.794570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.794580 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.811588 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.824449 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.840877 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.855208 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.875415 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:37Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.897512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.897580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.897603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.897638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4894]: I1210 18:56:37.897661 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.006319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.006380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.006394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.006412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.006425 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.109646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.109697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.109716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.109741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.109761 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.212348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.212395 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.212416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.212440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.212457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.315522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.315609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.315631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.315661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.315683 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.418758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.418817 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.418838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.418898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.418920 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.521526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.521590 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.521606 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.521633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.521652 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.540749 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.540769 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:38 crc kubenswrapper[4894]: E1210 18:56:38.541131 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:38 crc kubenswrapper[4894]: E1210 18:56:38.541266 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.624920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.624992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.625015 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.625042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.625057 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.727998 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.728051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.728067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.728090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.728109 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.831316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.831670 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.831970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.832238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.832461 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.935436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.935501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.935519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.935546 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4894]: I1210 18:56:38.935563 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.037966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.038012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.038028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.038052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.038068 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.141081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.141136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.141152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.141175 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.141193 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.245344 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.245411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.245433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.245464 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.245486 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.348214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.348271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.348287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.348310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.348326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.450605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.450669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.450686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.450710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.450728 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.540502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:39 crc kubenswrapper[4894]: E1210 18:56:39.540747 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.540825 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:39 crc kubenswrapper[4894]: E1210 18:56:39.541046 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.553167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.553256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.553275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.553300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.553317 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.656294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.656375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.656400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.656434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.656459 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.759670 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.759718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.759730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.759747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.759759 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.862636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.862709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.862741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.862758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.862770 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.964574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.964627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.964643 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.964664 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4894]: I1210 18:56:39.964678 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.078916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.078946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.078955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.078970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.078979 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.182388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.182423 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.182434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.182449 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.182461 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.237301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.237373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.237386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.237402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.237412 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.258226 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.261805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.261904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.261929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.261955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.261975 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.283029 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.288228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.288296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.288320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.288346 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.288364 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.309044 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.314050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.314091 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.314110 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.314136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.314154 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.331274 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.335935 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.335980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.335998 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.336022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.336040 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.364722 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c5920cce-fbb7-4806-aa5a-460529081e51\\\",\\\"systemUUID\\\":\\\"87dc1f21-a167-4b68-9d55-45806a47a7e3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.365360 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.367688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.367744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.367761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.367785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.367804 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.470779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.470934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.470952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.470976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.470997 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.540683 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.540901 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.540682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:40 crc kubenswrapper[4894]: E1210 18:56:40.541151 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.573672 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.573800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.573821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.573868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.573885 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.677020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.677079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.677096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.677118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.677136 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.780661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.780762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.780782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.780835 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.780911 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.883990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.884169 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.884193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.884218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.884268 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.987651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.987706 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.987731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.987760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4894]: I1210 18:56:40.987782 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.090530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.090582 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.090596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.090615 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.090627 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.193533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.193679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.193704 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.193732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.193754 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.296921 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.296979 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.296987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.297001 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.297011 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.445011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.445060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.445075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.445095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.445110 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.540743 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.540803 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:41 crc kubenswrapper[4894]: E1210 18:56:41.541030 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:41 crc kubenswrapper[4894]: E1210 18:56:41.541387 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.548072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.548139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.548161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.548187 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.548204 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.650458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.650560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.650577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.650604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.650622 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.753596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.753638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.753653 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.753673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.753686 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.856608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.856685 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.856709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.856742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.856766 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.960377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.960438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.960455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.960480 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4894]: I1210 18:56:41.960498 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.063527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.063603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.063623 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.063645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.063694 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.167337 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.167560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.167586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.167619 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.167642 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.271073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.271137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.271159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.271187 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.271207 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.374437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.374512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.374528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.374551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.374569 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.477405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.477475 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.477486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.477503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.477513 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.540790 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.540913 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:42 crc kubenswrapper[4894]: E1210 18:56:42.541008 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:42 crc kubenswrapper[4894]: E1210 18:56:42.541176 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.542617 4894 scope.go:117] "RemoveContainer" containerID="f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.580887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.580977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.580993 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.581016 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.581030 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.684779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.684828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.684840 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.684883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.684899 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.788576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.788628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.788645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.788668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.788687 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.892062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.892133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.892155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.892184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.892205 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.995684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.996093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.996290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.996503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4894]: I1210 18:56:42.996825 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.100404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.100443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.100454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.100471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.100487 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.203698 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.203771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.203789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.203816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.203834 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.307154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.307192 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.307208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.307226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.307237 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.410199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.410260 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.410276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.410296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.410310 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.513586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.513630 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.513643 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.513661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.513672 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.540001 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.540035 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:43 crc kubenswrapper[4894]: E1210 18:56:43.540162 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:43 crc kubenswrapper[4894]: E1210 18:56:43.540275 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.616709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.616754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.616768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.616788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.616800 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.718816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.718872 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.718880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.718895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.718904 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.822495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.822556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.822573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.822597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.822616 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.926214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.926279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.926302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.926327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4894]: I1210 18:56:43.926344 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.028399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.028490 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.028518 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.028549 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.028570 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.031292 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/2.log" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.035213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.035759 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.057983 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.074534 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.086502 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.104446 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.118158 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.131652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.131715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.131734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.131760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.131779 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.134667 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.146745 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.162521 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.181799 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.202154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.225210 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.234827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.234889 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.234899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.234913 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.234923 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.242423 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.256457 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.277816 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.296130 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.317740 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.330030 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.338162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.338405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.338590 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.338771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.338988 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.355134 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.442367 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.442431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.442448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.442474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.442491 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.540688 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.540700 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:44 crc kubenswrapper[4894]: E1210 18:56:44.540943 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:44 crc kubenswrapper[4894]: E1210 18:56:44.541128 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.546179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.546239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.546257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.546285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.546304 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.649764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.649885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.649913 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.649943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.649962 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.753229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.753282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.753298 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.753322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.753339 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.856349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.856686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.856818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.857066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.857211 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.960068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.960525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.960732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.960984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4894]: I1210 18:56:44.961194 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.041090 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/3.log" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.042148 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/2.log" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.047291 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" exitCode=1 Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.047356 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.047419 4894 scope.go:117] "RemoveContainer" containerID="f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.049752 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 18:56:45 crc kubenswrapper[4894]: E1210 18:56:45.050101 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.065587 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.066054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.066068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.066086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.066098 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.071045 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.089585 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.106172 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.119050 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.135433 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.153629 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.169296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.169377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.169401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.169429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.169446 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.169790 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.183180 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.198838 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.214245 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.230939 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.249066 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.267839 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.272268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.272316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.272335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.272359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.272375 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.279976 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.290784 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.304025 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.332693 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40e023af8ca4d5856d986da91a49ce9d66f207e3582ef97fc92e99d0769adb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:15Z\\\",\\\"message\\\":\\\".io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.512095 6592 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.513716 6592 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514220 6592 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 18:56:15.514496 6592 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:15.514768 6592 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:15.514808 6592 factory.go:656] Stopping watch factory\\\\nI1210 18:56:15.514820 6592 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:15.512500 6592 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-apiserver/api for endpointslice openshift-apiserver/api-qbcdq as it is not a known egress service\\\\nI1210 18:56:15.514872 6592 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:15.514899 6592 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:44Z\\\",\\\"message\\\":\\\"43.959674 7039 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:43.958912 7039 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:43.960280 7039 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:56:43.960307 7039 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:56:43.960343 7039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:56:43.960394 7039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:56:43.960425 7039 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:56:43.960547 7039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:43.960564 7039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 18:56:43.960569 7039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 18:56:43.960616 7039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:43.960631 7039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:56:43.960639 7039 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:56:43.960713 7039 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:56:43.960739 7039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.351389 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.376032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.376106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.376134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.376160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.376178 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.479436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.479508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.479530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.479560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.479582 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.540561 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.540561 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:45 crc kubenswrapper[4894]: E1210 18:56:45.540801 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:45 crc kubenswrapper[4894]: E1210 18:56:45.540960 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.583232 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.583301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.583324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.583353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.583373 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.686369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.686416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.686471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.686493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.686509 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.789946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.790001 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.790019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.790044 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.790067 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.892701 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.892750 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.892762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.892779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.892792 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.996682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.996743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.996764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.996789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4894]: I1210 18:56:45.996809 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.055692 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/3.log" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.061492 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.061757 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.083341 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.095749 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.095971 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:50.095938717 +0000 UTC m=+152.890786549 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.099813 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.099909 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.099933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.099958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.099977 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.103115 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.137258 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:44Z\\\",\\\"message\\\":\\\"43.959674 7039 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:43.958912 7039 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:43.960280 7039 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:56:43.960307 7039 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:56:43.960343 7039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:56:43.960394 7039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:56:43.960425 7039 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:56:43.960547 7039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:43.960564 7039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 18:56:43.960569 7039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 18:56:43.960616 7039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:43.960631 7039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:56:43.960639 7039 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:56:43.960713 7039 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:56:43.960739 7039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.156160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.172778 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.188674 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.197300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.197342 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.197395 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.197421 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197506 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197552 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:50.197538186 +0000 UTC m=+152.992385988 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197555 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197609 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197652 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197670 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:50.197638819 +0000 UTC m=+152.992486651 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197676 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197753 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:50.197728711 +0000 UTC m=+152.992576543 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197758 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197790 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197806 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.197874 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:50.197833704 +0000 UTC m=+152.992681546 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.203475 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.203517 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.203530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.203550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.203562 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.205702 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.226908 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.247969 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.269285 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.289886 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.304037 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.307183 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.307240 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.307257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.307282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.307300 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.324229 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.342703 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.362680 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.381808 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.397328 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.409800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.409826 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.409837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.409875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.409890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.418720 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.512679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.512734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.512751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.512773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.512791 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.540272 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.540373 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.540465 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:46 crc kubenswrapper[4894]: E1210 18:56:46.540609 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.615338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.615394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.615414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.615438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.615454 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.717962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.718031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.718052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.718079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.718099 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.820974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.821035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.821052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.821078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.821097 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.923772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.923901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.923936 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.923962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4894]: I1210 18:56:46.923980 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.026991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.027019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.027027 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.027040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.027048 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.129931 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.129996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.130010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.130037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.130053 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.232449 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.232500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.232512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.232531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.232542 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.335136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.335200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.335217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.335246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.335297 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.439143 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.439215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.439230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.439264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.439280 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.540543 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.540634 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:47 crc kubenswrapper[4894]: E1210 18:56:47.540736 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:47 crc kubenswrapper[4894]: E1210 18:56:47.540815 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.545503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.545556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.545578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.545606 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.545628 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.563077 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb3fbf28-4903-4e44-8144-86c98da53f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:31.439765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:31.441706 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2497523109/tls.crt::/tmp/serving-cert-2497523109/tls.key\\\\\\\"\\\\nI1210 18:55:42.203781 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:42.210276 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:42.210316 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:42.210373 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:42.210381 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:42.217316 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:42.217347 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:42.217354 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217366 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:42.217375 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:42.217380 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:42.217385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:42.217391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:42.221230 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.581724 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0bfcea8f5cf510403a6149e94dd41f545465717e3d9fec62da28b17f0d0c257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4e6e4b8b24dc22ff580a77b1837124e9d944c560a89646f686d1e30056013f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.593504 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n6rc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f663d56-37e2-4a3f-9bb1-24e31b011470\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a33775bd99467a3608a91f019e216e6c016f0779cfe2eb168eb92aab0b9e857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czrsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n6rc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.605549 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"608bd9cb-a410-43f0-b856-0b9804087a90\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83161618186f4e1959b690a404fa7fbac20b44f92d08a8f353cdba6230a6bf7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://560dd57e4089fb3b2ff93f2b89143680ba0383f36ebf32aa5500664612c64a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c37e4fa300e5c7932a5cd804f8307ca2546ec806ce3ca90afbee01c42bb2783f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76e701488a005a536ddcbe3d40fc0c4fee3db1b416daf73927380bca728bb75b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.623611 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da5950e6d74880d2167b8c1d26f603c7ba7740d41a964229390f914735a5f5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.646542 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.650400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.650544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.650558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.650578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.650591 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.666647 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-msfl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9392e0e-ca87-437a-8ba1-9c5d34a50660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:35Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d\\\\n2025-12-10T18:55:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_945d2f52-2f92-4754-ba84-36347635ba6d to /host/opt/cni/bin/\\\\n2025-12-10T18:55:50Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:50Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-64fjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-msfl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.683588 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d85c5b7-37ba-4aaa-a1d9-52882a94e0ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4536d0b319b6f982991d602c8d77c35583b782e4643340fad26b85321f374bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49c107f43462d44471bc0c67874e7849e6b9029167626604ffbee289a39c3cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2vwtx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxsrp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.698906 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dks9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72521f10-d803-4b10-97dc-cbad9bda6be7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8szbg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:56:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dks9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.713682 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e005f2f2-fa5d-459b-bc5c-667bbab1315e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://317a0f622658d762d826ffce8531a85c519676c33215b5c09f85d507d3a181dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45b156ce029ce4cfbb6f9514742d5be4cb2e5adba7faccad382f33c98cedee78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.731241 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d18a0ac9bde9382c447fca40d8d963ee14c3a1a9d4c6aca6f168ae7a391bd04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.751192 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.752793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.752834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.752889 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.752911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.752927 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.766659 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.782538 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb325131-20c4-4791-9bbc-45ae36443184\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a42bab541ad7ff271106eeb3a0b50ca0504f786202d87b3e2ae2efd5db5e957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6828160a4952755f2a1cdf258fa5cbf2bb0831aed8a9b4eb5f977d649472536\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14574f7c45a4fbfd9e9c8f0fbd0abfc9b026238624253f0824437639c7488180\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5c0325383af9e1c25231412dca8872dd1841eee662ae9c883361edc54fea77\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be709083a0baec6afe0de1524eb5584cdab020afa2b818b064ffdece70e51378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e7de533f968862dafd0009fb8d6c7756ce0eaaa3b8718c0685db7ad78e2adf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0a2fb61d9cd2e059e15d4d61882aa5f2ed96385bb5b6c4ff168400d27d6d6fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pvr6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-q8rg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.797639 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fd55194-d263-4547-9b8f-934ef8050b59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b66f67f4f0317365a4a84a2a43d49b1bf1a0f5fa3812882b64813b5e80fb718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5ftsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-vbjj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.814003 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e13cf33-d0ed-4ee4-ad23-8fc2373d3ceb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9092e4d285678afe06cf82e7d4be9d22a03b2692b01b591a3b932676edee56cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa195ab6bfc881689c4bf63911dd95e3ecccd885f65d2ce10a0490fe1b043027\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://47bb118a18d229c63c9181c9767c43920f4b1dc9b8acabebde9ee15c294c154b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.825759 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wnfrc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c842b50-bc0f-47a6-97aa-67c9562005aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f1d5e216ae49878e67945b7a3423293fd7f22cf2f8d1bdebdc62bb33b57cb45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szhgj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wnfrc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.847299 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de31cb9-e20f-4578-b0ac-b03df0bc133c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:44Z\\\",\\\"message\\\":\\\"43.959674 7039 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:43.958912 7039 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:56:43.960280 7039 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:56:43.960307 7039 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:56:43.960343 7039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:56:43.960394 7039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:56:43.960425 7039 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:56:43.960547 7039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 18:56:43.960564 7039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 18:56:43.960569 7039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 18:56:43.960616 7039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:56:43.960631 7039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1210 18:56:43.960639 7039 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:56:43.960713 7039 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:56:43.960739 7039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bnqbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jqwjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.855762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.855799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.855814 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.855835 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.855877 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.957715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.957999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.958085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.958167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4894]: I1210 18:56:47.958250 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.060308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.060356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.060368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.060384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.060396 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.163380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.163703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.163838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.164125 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.164266 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.267043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.267124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.267148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.267179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.267202 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.373061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.373541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.373790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.374063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.374284 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.478219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.478280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.478301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.478326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.478344 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.540667 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.540669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:48 crc kubenswrapper[4894]: E1210 18:56:48.540905 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:48 crc kubenswrapper[4894]: E1210 18:56:48.541010 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.581439 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.581502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.581520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.581543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.581561 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.684175 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.684598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.684758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.684937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.685172 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.787466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.787522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.787556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.787589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.787603 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.890202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.890262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.890279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.890303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.890323 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.993454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.993497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.993505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.993519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4894]: I1210 18:56:48.993529 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.096873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.096938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.096956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.096982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.097000 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.200987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.201050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.201063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.201084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.201099 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.304126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.304152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.304160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.304171 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.304180 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.406000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.406043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.406055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.406069 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.406078 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.508159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.508206 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.508216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.508244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.508257 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.540295 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.540384 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:49 crc kubenswrapper[4894]: E1210 18:56:49.540460 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:49 crc kubenswrapper[4894]: E1210 18:56:49.540556 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.611010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.611055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.611064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.611078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.611088 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.714253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.714304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.714319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.714340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.714354 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.817648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.817713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.817738 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.817766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.817787 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.920718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.920832 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.920895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.920937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4894]: I1210 18:56:49.920964 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.023663 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.023737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.023764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.023794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.023815 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.126064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.126134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.126150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.126187 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.126217 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.229281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.229331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.229350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.229373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.229391 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.332176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.332260 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.332279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.332307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.332327 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.434900 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.434959 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.434976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.435000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.435023 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.537366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.537421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.537438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.537462 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.537479 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.539812 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.539893 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:50 crc kubenswrapper[4894]: E1210 18:56:50.540007 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:50 crc kubenswrapper[4894]: E1210 18:56:50.540147 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.640400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.640478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.640498 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.640525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.640544 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.741402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.741472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.741491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.741525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.741550 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.790079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.790119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.790129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.790146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.790157 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.815577 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp"] Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.816229 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.818685 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.820757 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.821811 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.821956 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.851077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/137f9100-9373-4cea-9f40-b843ad6c0902-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.851172 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/137f9100-9373-4cea-9f40-b843ad6c0902-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.851309 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/137f9100-9373-4cea-9f40-b843ad6c0902-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.851390 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/137f9100-9373-4cea-9f40-b843ad6c0902-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.851462 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/137f9100-9373-4cea-9f40-b843ad6c0902-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.861678 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.861656244 podStartE2EDuration="43.861656244s" podCreationTimestamp="2025-12-10 18:56:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.847761071 +0000 UTC m=+93.642608863" watchObservedRunningTime="2025-12-10 18:56:50.861656244 +0000 UTC m=+93.656504026" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.891364 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-msfl4" podStartSLOduration=62.891345181 podStartE2EDuration="1m2.891345181s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.890970851 +0000 UTC m=+93.685818663" watchObservedRunningTime="2025-12-10 18:56:50.891345181 +0000 UTC m=+93.686192973" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.908130 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxsrp" podStartSLOduration=62.908105941 podStartE2EDuration="1m2.908105941s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.906992721 +0000 UTC m=+93.701840523" watchObservedRunningTime="2025-12-10 18:56:50.908105941 +0000 UTC m=+93.702953753" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.941335 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podStartSLOduration=62.941311793 podStartE2EDuration="1m2.941311793s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.941114788 +0000 UTC m=+93.735962620" watchObservedRunningTime="2025-12-10 18:56:50.941311793 +0000 UTC m=+93.736159585" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.941663 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-q8rg7" podStartSLOduration=62.941656682 podStartE2EDuration="1m2.941656682s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.926924076 +0000 UTC m=+93.721771868" watchObservedRunningTime="2025-12-10 18:56:50.941656682 +0000 UTC m=+93.736504464" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.950946 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.950924571 podStartE2EDuration="21.950924571s" podCreationTimestamp="2025-12-10 18:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.950654243 +0000 UTC m=+93.745502055" watchObservedRunningTime="2025-12-10 18:56:50.950924571 +0000 UTC m=+93.745772373" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953072 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/137f9100-9373-4cea-9f40-b843ad6c0902-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953143 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/137f9100-9373-4cea-9f40-b843ad6c0902-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953174 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/137f9100-9373-4cea-9f40-b843ad6c0902-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953196 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/137f9100-9373-4cea-9f40-b843ad6c0902-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/137f9100-9373-4cea-9f40-b843ad6c0902-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953241 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/137f9100-9373-4cea-9f40-b843ad6c0902-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.953353 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/137f9100-9373-4cea-9f40-b843ad6c0902-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.954374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/137f9100-9373-4cea-9f40-b843ad6c0902-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.961061 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/137f9100-9373-4cea-9f40-b843ad6c0902-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:50 crc kubenswrapper[4894]: I1210 18:56:50.980332 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/137f9100-9373-4cea-9f40-b843ad6c0902-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j6mhp\" (UID: \"137f9100-9373-4cea-9f40-b843ad6c0902\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.014879 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.014831137 podStartE2EDuration="1m7.014831137s" podCreationTimestamp="2025-12-10 18:55:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:51.014301883 +0000 UTC m=+93.809149675" watchObservedRunningTime="2025-12-10 18:56:51.014831137 +0000 UTC m=+93.809678929" Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.025474 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-wnfrc" podStartSLOduration=63.025456763 podStartE2EDuration="1m3.025456763s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:51.025420172 +0000 UTC m=+93.820267964" watchObservedRunningTime="2025-12-10 18:56:51.025456763 +0000 UTC m=+93.820304555" Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.113544 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.113521377 podStartE2EDuration="1m9.113521377s" podCreationTimestamp="2025-12-10 18:55:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:51.09054356 +0000 UTC m=+93.885391372" watchObservedRunningTime="2025-12-10 18:56:51.113521377 +0000 UTC m=+93.908369179" Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.134380 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.140304 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-n6rc2" podStartSLOduration=63.140286856 podStartE2EDuration="1m3.140286856s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:51.128257193 +0000 UTC m=+93.923104995" watchObservedRunningTime="2025-12-10 18:56:51.140286856 +0000 UTC m=+93.935134648" Dec 10 18:56:51 crc kubenswrapper[4894]: W1210 18:56:51.150344 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod137f9100_9373_4cea_9f40_b843ad6c0902.slice/crio-9bfe5e2b07eb7efd0603eacccd5681a091fccbcf21e52f70b5d99b1e16d9672e WatchSource:0}: Error finding container 9bfe5e2b07eb7efd0603eacccd5681a091fccbcf21e52f70b5d99b1e16d9672e: Status 404 returned error can't find the container with id 9bfe5e2b07eb7efd0603eacccd5681a091fccbcf21e52f70b5d99b1e16d9672e Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.540469 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:51 crc kubenswrapper[4894]: E1210 18:56:51.540949 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:51 crc kubenswrapper[4894]: I1210 18:56:51.540729 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:51 crc kubenswrapper[4894]: E1210 18:56:51.541246 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:52 crc kubenswrapper[4894]: I1210 18:56:52.081078 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" event={"ID":"137f9100-9373-4cea-9f40-b843ad6c0902","Type":"ContainerStarted","Data":"3b6c2ab309585ca180dcc9722731522823d421d296a4f1dd9d2984951461a2b3"} Dec 10 18:56:52 crc kubenswrapper[4894]: I1210 18:56:52.081138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" event={"ID":"137f9100-9373-4cea-9f40-b843ad6c0902","Type":"ContainerStarted","Data":"9bfe5e2b07eb7efd0603eacccd5681a091fccbcf21e52f70b5d99b1e16d9672e"} Dec 10 18:56:52 crc kubenswrapper[4894]: I1210 18:56:52.098540 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6mhp" podStartSLOduration=64.098519108 podStartE2EDuration="1m4.098519108s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:52.097748477 +0000 UTC m=+94.892596289" watchObservedRunningTime="2025-12-10 18:56:52.098519108 +0000 UTC m=+94.893366910" Dec 10 18:56:52 crc kubenswrapper[4894]: I1210 18:56:52.539963 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:52 crc kubenswrapper[4894]: I1210 18:56:52.539982 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:52 crc kubenswrapper[4894]: E1210 18:56:52.540178 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:52 crc kubenswrapper[4894]: E1210 18:56:52.540343 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:53 crc kubenswrapper[4894]: I1210 18:56:53.540952 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:53 crc kubenswrapper[4894]: E1210 18:56:53.541079 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:53 crc kubenswrapper[4894]: I1210 18:56:53.541243 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:53 crc kubenswrapper[4894]: E1210 18:56:53.541326 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:54 crc kubenswrapper[4894]: I1210 18:56:54.539982 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:54 crc kubenswrapper[4894]: I1210 18:56:54.540092 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:54 crc kubenswrapper[4894]: E1210 18:56:54.540216 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:54 crc kubenswrapper[4894]: E1210 18:56:54.540730 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:54 crc kubenswrapper[4894]: I1210 18:56:54.568362 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 10 18:56:55 crc kubenswrapper[4894]: I1210 18:56:55.540594 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:55 crc kubenswrapper[4894]: E1210 18:56:55.540709 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:55 crc kubenswrapper[4894]: I1210 18:56:55.540600 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:55 crc kubenswrapper[4894]: E1210 18:56:55.540988 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:56 crc kubenswrapper[4894]: I1210 18:56:56.540661 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:56 crc kubenswrapper[4894]: I1210 18:56:56.540708 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:56 crc kubenswrapper[4894]: E1210 18:56:56.541036 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:56 crc kubenswrapper[4894]: E1210 18:56:56.541168 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:57 crc kubenswrapper[4894]: I1210 18:56:57.541121 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:57 crc kubenswrapper[4894]: E1210 18:56:57.542473 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:57 crc kubenswrapper[4894]: I1210 18:56:57.542789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:57 crc kubenswrapper[4894]: E1210 18:56:57.542972 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:58 crc kubenswrapper[4894]: I1210 18:56:58.540184 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:56:58 crc kubenswrapper[4894]: I1210 18:56:58.540184 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:58 crc kubenswrapper[4894]: E1210 18:56:58.540398 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:56:58 crc kubenswrapper[4894]: E1210 18:56:58.540546 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:58 crc kubenswrapper[4894]: I1210 18:56:58.542351 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 18:56:58 crc kubenswrapper[4894]: E1210 18:56:58.542661 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:56:59 crc kubenswrapper[4894]: I1210 18:56:59.540789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:59 crc kubenswrapper[4894]: I1210 18:56:59.540792 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:59 crc kubenswrapper[4894]: E1210 18:56:59.541246 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:59 crc kubenswrapper[4894]: E1210 18:56:59.541255 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:00 crc kubenswrapper[4894]: I1210 18:57:00.540209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:00 crc kubenswrapper[4894]: I1210 18:57:00.540310 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:00 crc kubenswrapper[4894]: E1210 18:57:00.540739 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:00 crc kubenswrapper[4894]: E1210 18:57:00.540902 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:01 crc kubenswrapper[4894]: I1210 18:57:01.540003 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:01 crc kubenswrapper[4894]: I1210 18:57:01.540111 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:01 crc kubenswrapper[4894]: E1210 18:57:01.540181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:01 crc kubenswrapper[4894]: E1210 18:57:01.540272 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:02 crc kubenswrapper[4894]: I1210 18:57:02.540965 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:02 crc kubenswrapper[4894]: I1210 18:57:02.541126 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:02 crc kubenswrapper[4894]: E1210 18:57:02.541487 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:02 crc kubenswrapper[4894]: E1210 18:57:02.541663 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:03 crc kubenswrapper[4894]: I1210 18:57:03.540208 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:03 crc kubenswrapper[4894]: I1210 18:57:03.540208 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:03 crc kubenswrapper[4894]: E1210 18:57:03.540342 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:03 crc kubenswrapper[4894]: E1210 18:57:03.540441 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:04 crc kubenswrapper[4894]: I1210 18:57:04.540452 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:04 crc kubenswrapper[4894]: I1210 18:57:04.540553 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:04 crc kubenswrapper[4894]: E1210 18:57:04.540631 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:04 crc kubenswrapper[4894]: E1210 18:57:04.540728 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:05 crc kubenswrapper[4894]: I1210 18:57:05.540753 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:05 crc kubenswrapper[4894]: E1210 18:57:05.541003 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:05 crc kubenswrapper[4894]: I1210 18:57:05.541066 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:05 crc kubenswrapper[4894]: E1210 18:57:05.541262 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:06 crc kubenswrapper[4894]: I1210 18:57:06.540340 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:06 crc kubenswrapper[4894]: E1210 18:57:06.540499 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:06 crc kubenswrapper[4894]: I1210 18:57:06.540344 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:06 crc kubenswrapper[4894]: E1210 18:57:06.540702 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:07 crc kubenswrapper[4894]: I1210 18:57:07.032800 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:07 crc kubenswrapper[4894]: E1210 18:57:07.033013 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:57:07 crc kubenswrapper[4894]: E1210 18:57:07.033079 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs podName:72521f10-d803-4b10-97dc-cbad9bda6be7 nodeName:}" failed. No retries permitted until 2025-12-10 18:58:11.033063113 +0000 UTC m=+173.827910905 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs") pod "network-metrics-daemon-7dks9" (UID: "72521f10-d803-4b10-97dc-cbad9bda6be7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:57:07 crc kubenswrapper[4894]: I1210 18:57:07.540639 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:07 crc kubenswrapper[4894]: I1210 18:57:07.540711 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:07 crc kubenswrapper[4894]: E1210 18:57:07.542540 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:07 crc kubenswrapper[4894]: E1210 18:57:07.542911 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:08 crc kubenswrapper[4894]: I1210 18:57:08.540268 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:08 crc kubenswrapper[4894]: I1210 18:57:08.540268 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:08 crc kubenswrapper[4894]: E1210 18:57:08.540454 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:08 crc kubenswrapper[4894]: E1210 18:57:08.540582 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:09 crc kubenswrapper[4894]: I1210 18:57:09.539914 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:09 crc kubenswrapper[4894]: I1210 18:57:09.539980 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:09 crc kubenswrapper[4894]: E1210 18:57:09.540047 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:09 crc kubenswrapper[4894]: E1210 18:57:09.540213 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:10 crc kubenswrapper[4894]: I1210 18:57:10.540146 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:10 crc kubenswrapper[4894]: I1210 18:57:10.540195 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:10 crc kubenswrapper[4894]: E1210 18:57:10.541183 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:10 crc kubenswrapper[4894]: E1210 18:57:10.541257 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:11 crc kubenswrapper[4894]: I1210 18:57:11.540024 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:11 crc kubenswrapper[4894]: I1210 18:57:11.540708 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:11 crc kubenswrapper[4894]: E1210 18:57:11.540783 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:11 crc kubenswrapper[4894]: E1210 18:57:11.541941 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:11 crc kubenswrapper[4894]: I1210 18:57:11.542277 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 18:57:11 crc kubenswrapper[4894]: E1210 18:57:11.543213 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:57:12 crc kubenswrapper[4894]: I1210 18:57:12.540884 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:12 crc kubenswrapper[4894]: I1210 18:57:12.540994 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:12 crc kubenswrapper[4894]: E1210 18:57:12.541075 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:12 crc kubenswrapper[4894]: E1210 18:57:12.541173 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:13 crc kubenswrapper[4894]: I1210 18:57:13.540682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:13 crc kubenswrapper[4894]: I1210 18:57:13.540800 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:13 crc kubenswrapper[4894]: E1210 18:57:13.540943 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:13 crc kubenswrapper[4894]: E1210 18:57:13.541027 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:14 crc kubenswrapper[4894]: I1210 18:57:14.539914 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:14 crc kubenswrapper[4894]: E1210 18:57:14.540107 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:14 crc kubenswrapper[4894]: I1210 18:57:14.539943 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:14 crc kubenswrapper[4894]: E1210 18:57:14.540281 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:15 crc kubenswrapper[4894]: I1210 18:57:15.539918 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:15 crc kubenswrapper[4894]: I1210 18:57:15.539953 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:15 crc kubenswrapper[4894]: E1210 18:57:15.540101 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:15 crc kubenswrapper[4894]: E1210 18:57:15.540245 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:16 crc kubenswrapper[4894]: I1210 18:57:16.540921 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:16 crc kubenswrapper[4894]: I1210 18:57:16.540947 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:16 crc kubenswrapper[4894]: E1210 18:57:16.541134 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:16 crc kubenswrapper[4894]: E1210 18:57:16.541223 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:17 crc kubenswrapper[4894]: E1210 18:57:17.517032 4894 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 10 18:57:17 crc kubenswrapper[4894]: I1210 18:57:17.540110 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:17 crc kubenswrapper[4894]: E1210 18:57:17.542805 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:17 crc kubenswrapper[4894]: I1210 18:57:17.542947 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:17 crc kubenswrapper[4894]: E1210 18:57:17.543193 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:17 crc kubenswrapper[4894]: E1210 18:57:17.629346 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:18 crc kubenswrapper[4894]: I1210 18:57:18.540896 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:18 crc kubenswrapper[4894]: I1210 18:57:18.540938 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:18 crc kubenswrapper[4894]: E1210 18:57:18.541399 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:18 crc kubenswrapper[4894]: E1210 18:57:18.541576 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:19 crc kubenswrapper[4894]: I1210 18:57:19.540715 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:19 crc kubenswrapper[4894]: E1210 18:57:19.540929 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:19 crc kubenswrapper[4894]: I1210 18:57:19.541041 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:19 crc kubenswrapper[4894]: E1210 18:57:19.541245 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:20 crc kubenswrapper[4894]: I1210 18:57:20.540382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:20 crc kubenswrapper[4894]: E1210 18:57:20.540558 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:20 crc kubenswrapper[4894]: I1210 18:57:20.540382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:20 crc kubenswrapper[4894]: E1210 18:57:20.540719 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:21 crc kubenswrapper[4894]: I1210 18:57:21.540163 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:21 crc kubenswrapper[4894]: I1210 18:57:21.540219 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:21 crc kubenswrapper[4894]: E1210 18:57:21.540307 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:21 crc kubenswrapper[4894]: E1210 18:57:21.540365 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.185591 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/1.log" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.186398 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/0.log" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.186442 4894 generic.go:334] "Generic (PLEG): container finished" podID="d9392e0e-ca87-437a-8ba1-9c5d34a50660" containerID="dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53" exitCode=1 Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.186471 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerDied","Data":"dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53"} Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.186502 4894 scope.go:117] "RemoveContainer" containerID="a33fee19c6ec2f42d250d9343cb88080170020b61ea9c1c766e93b9b90ded82d" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.186946 4894 scope.go:117] "RemoveContainer" containerID="dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53" Dec 10 18:57:22 crc kubenswrapper[4894]: E1210 18:57:22.187113 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-msfl4_openshift-multus(d9392e0e-ca87-437a-8ba1-9c5d34a50660)\"" pod="openshift-multus/multus-msfl4" podUID="d9392e0e-ca87-437a-8ba1-9c5d34a50660" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.206720 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=28.206700954 podStartE2EDuration="28.206700954s" podCreationTimestamp="2025-12-10 18:56:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:57.581430582 +0000 UTC m=+100.376278434" watchObservedRunningTime="2025-12-10 18:57:22.206700954 +0000 UTC m=+125.001548746" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.541002 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.541067 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:22 crc kubenswrapper[4894]: E1210 18:57:22.541252 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:22 crc kubenswrapper[4894]: E1210 18:57:22.541522 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:22 crc kubenswrapper[4894]: I1210 18:57:22.542993 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 18:57:22 crc kubenswrapper[4894]: E1210 18:57:22.543303 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jqwjd_openshift-ovn-kubernetes(8de31cb9-e20f-4578-b0ac-b03df0bc133c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" Dec 10 18:57:22 crc kubenswrapper[4894]: E1210 18:57:22.631178 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:23 crc kubenswrapper[4894]: I1210 18:57:23.193173 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/1.log" Dec 10 18:57:23 crc kubenswrapper[4894]: I1210 18:57:23.540127 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:23 crc kubenswrapper[4894]: I1210 18:57:23.540230 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:23 crc kubenswrapper[4894]: E1210 18:57:23.540270 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:23 crc kubenswrapper[4894]: E1210 18:57:23.540442 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:24 crc kubenswrapper[4894]: I1210 18:57:24.540196 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:24 crc kubenswrapper[4894]: I1210 18:57:24.540217 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:24 crc kubenswrapper[4894]: E1210 18:57:24.541119 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:24 crc kubenswrapper[4894]: E1210 18:57:24.541344 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:25 crc kubenswrapper[4894]: I1210 18:57:25.540831 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:25 crc kubenswrapper[4894]: I1210 18:57:25.541204 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:25 crc kubenswrapper[4894]: E1210 18:57:25.541373 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:25 crc kubenswrapper[4894]: E1210 18:57:25.541574 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:26 crc kubenswrapper[4894]: I1210 18:57:26.540502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:26 crc kubenswrapper[4894]: I1210 18:57:26.540520 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:26 crc kubenswrapper[4894]: E1210 18:57:26.540693 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:26 crc kubenswrapper[4894]: E1210 18:57:26.540801 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:27 crc kubenswrapper[4894]: I1210 18:57:27.540783 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:27 crc kubenswrapper[4894]: E1210 18:57:27.543159 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:27 crc kubenswrapper[4894]: I1210 18:57:27.543261 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:27 crc kubenswrapper[4894]: E1210 18:57:27.544266 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:27 crc kubenswrapper[4894]: E1210 18:57:27.631694 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:28 crc kubenswrapper[4894]: I1210 18:57:28.539953 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:28 crc kubenswrapper[4894]: I1210 18:57:28.540020 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:28 crc kubenswrapper[4894]: E1210 18:57:28.540189 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:28 crc kubenswrapper[4894]: E1210 18:57:28.540280 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:29 crc kubenswrapper[4894]: I1210 18:57:29.541058 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:29 crc kubenswrapper[4894]: I1210 18:57:29.541095 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:29 crc kubenswrapper[4894]: E1210 18:57:29.541286 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:29 crc kubenswrapper[4894]: E1210 18:57:29.541387 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:30 crc kubenswrapper[4894]: I1210 18:57:30.540139 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:30 crc kubenswrapper[4894]: E1210 18:57:30.540336 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:30 crc kubenswrapper[4894]: I1210 18:57:30.540620 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:30 crc kubenswrapper[4894]: E1210 18:57:30.541016 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:31 crc kubenswrapper[4894]: I1210 18:57:31.540671 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:31 crc kubenswrapper[4894]: I1210 18:57:31.540791 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:31 crc kubenswrapper[4894]: E1210 18:57:31.540916 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:31 crc kubenswrapper[4894]: E1210 18:57:31.541146 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:32 crc kubenswrapper[4894]: I1210 18:57:32.540545 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:32 crc kubenswrapper[4894]: I1210 18:57:32.540592 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:32 crc kubenswrapper[4894]: E1210 18:57:32.540678 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:32 crc kubenswrapper[4894]: E1210 18:57:32.540799 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:32 crc kubenswrapper[4894]: E1210 18:57:32.633374 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:33 crc kubenswrapper[4894]: I1210 18:57:33.540559 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:33 crc kubenswrapper[4894]: I1210 18:57:33.540594 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:33 crc kubenswrapper[4894]: E1210 18:57:33.541935 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:33 crc kubenswrapper[4894]: E1210 18:57:33.542123 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:34 crc kubenswrapper[4894]: I1210 18:57:34.540477 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:34 crc kubenswrapper[4894]: I1210 18:57:34.540474 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:34 crc kubenswrapper[4894]: E1210 18:57:34.540683 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:34 crc kubenswrapper[4894]: E1210 18:57:34.540954 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:34 crc kubenswrapper[4894]: I1210 18:57:34.542091 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.240682 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/3.log" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.243338 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerStarted","Data":"8e5d9ff95576272fc324ba39a9ab49fcad475765ccacc2d6184b5302bbbf7ff8"} Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.243885 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.280048 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podStartSLOduration=107.280025426 podStartE2EDuration="1m47.280025426s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:35.278377564 +0000 UTC m=+138.073225376" watchObservedRunningTime="2025-12-10 18:57:35.280025426 +0000 UTC m=+138.074873248" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.540979 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:35 crc kubenswrapper[4894]: E1210 18:57:35.541370 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.541392 4894 scope.go:117] "RemoveContainer" containerID="dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.541473 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:35 crc kubenswrapper[4894]: E1210 18:57:35.541569 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.555519 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7dks9"] Dec 10 18:57:35 crc kubenswrapper[4894]: I1210 18:57:35.555633 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:35 crc kubenswrapper[4894]: E1210 18:57:35.555760 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:36 crc kubenswrapper[4894]: I1210 18:57:36.248440 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/1.log" Dec 10 18:57:36 crc kubenswrapper[4894]: I1210 18:57:36.248822 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerStarted","Data":"d422371c764f725f4e201df30c379baf3507de97186b2eeae087e857e84a3523"} Dec 10 18:57:36 crc kubenswrapper[4894]: I1210 18:57:36.540500 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:36 crc kubenswrapper[4894]: E1210 18:57:36.540715 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:37 crc kubenswrapper[4894]: I1210 18:57:37.539980 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:37 crc kubenswrapper[4894]: I1210 18:57:37.539990 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:37 crc kubenswrapper[4894]: I1210 18:57:37.540105 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:37 crc kubenswrapper[4894]: E1210 18:57:37.541415 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:37 crc kubenswrapper[4894]: E1210 18:57:37.541539 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:37 crc kubenswrapper[4894]: E1210 18:57:37.541696 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:37 crc kubenswrapper[4894]: E1210 18:57:37.634146 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:38 crc kubenswrapper[4894]: I1210 18:57:38.540222 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:38 crc kubenswrapper[4894]: E1210 18:57:38.540343 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:39 crc kubenswrapper[4894]: I1210 18:57:39.540420 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:39 crc kubenswrapper[4894]: I1210 18:57:39.540471 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:39 crc kubenswrapper[4894]: I1210 18:57:39.540514 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:39 crc kubenswrapper[4894]: E1210 18:57:39.540644 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:39 crc kubenswrapper[4894]: E1210 18:57:39.540790 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:39 crc kubenswrapper[4894]: E1210 18:57:39.541013 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:40 crc kubenswrapper[4894]: I1210 18:57:40.540925 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:40 crc kubenswrapper[4894]: E1210 18:57:40.541127 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:41 crc kubenswrapper[4894]: I1210 18:57:41.540286 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:41 crc kubenswrapper[4894]: I1210 18:57:41.540425 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:41 crc kubenswrapper[4894]: E1210 18:57:41.540464 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dks9" podUID="72521f10-d803-4b10-97dc-cbad9bda6be7" Dec 10 18:57:41 crc kubenswrapper[4894]: E1210 18:57:41.540628 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:41 crc kubenswrapper[4894]: I1210 18:57:41.540938 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:41 crc kubenswrapper[4894]: E1210 18:57:41.541054 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:42 crc kubenswrapper[4894]: I1210 18:57:42.539962 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:42 crc kubenswrapper[4894]: E1210 18:57:42.540138 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.540947 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.540968 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.541347 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.544228 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.544379 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.544427 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.544988 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.546643 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 18:57:43 crc kubenswrapper[4894]: I1210 18:57:43.547079 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 18:57:44 crc kubenswrapper[4894]: I1210 18:57:44.540919 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:49 crc kubenswrapper[4894]: I1210 18:57:49.577510 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 18:57:49 crc kubenswrapper[4894]: I1210 18:57:49.579235 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.045224 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.132568 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:50 crc kubenswrapper[4894]: E1210 18:57:50.132947 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:59:52.132902412 +0000 UTC m=+274.927750244 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.234256 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.234316 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.234348 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.234388 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.235946 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.240923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.241397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.248509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.263098 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.468027 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:50 crc kubenswrapper[4894]: I1210 18:57:50.499014 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:50 crc kubenswrapper[4894]: W1210 18:57:50.668662 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e0ae8b269cd22322acb08897e88715818bbc03ef68fb12dbaa67868f91962c0b WatchSource:0}: Error finding container e0ae8b269cd22322acb08897e88715818bbc03ef68fb12dbaa67868f91962c0b: Status 404 returned error can't find the container with id e0ae8b269cd22322acb08897e88715818bbc03ef68fb12dbaa67868f91962c0b Dec 10 18:57:50 crc kubenswrapper[4894]: W1210 18:57:50.709078 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-4414a42ba33a737cbabdb866cf53c4bd5403663758a5976524a6511db5162c10 WatchSource:0}: Error finding container 4414a42ba33a737cbabdb866cf53c4bd5403663758a5976524a6511db5162c10: Status 404 returned error can't find the container with id 4414a42ba33a737cbabdb866cf53c4bd5403663758a5976524a6511db5162c10 Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.306911 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e0ae8b269cd22322acb08897e88715818bbc03ef68fb12dbaa67868f91962c0b"} Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.309210 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0aec43a2799a63bc557f21d50490e2f5b314a84cceeced836c84bae54912d537"} Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.311576 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4414a42ba33a737cbabdb866cf53c4bd5403663758a5976524a6511db5162c10"} Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.581379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.639666 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bl258"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.640507 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k47sq"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.641127 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.641334 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vsms7"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.641748 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.641820 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.644028 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6pnv8"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.645230 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.646147 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.647153 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.648514 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.652430 4894 reflector.go:561] object-"openshift-console"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.652503 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.654653 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.656110 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m2s4d"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.656802 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dwxt5"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.657376 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.657489 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.657629 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.658156 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.661458 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bqdvr"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.662267 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.667081 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.667316 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.667474 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.667978 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.668137 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.675307 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.675663 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.675942 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.676284 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.676333 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.676675 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.677376 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.677424 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.677594 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.677822 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.676293 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678144 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.678271 4894 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-session": failed to list *v1.Secret: secrets "v4-0-config-system-session" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678287 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.678298 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-session\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-session\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.678406 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.678426 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678431 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678485 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.678514 4894 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-provider-selection": failed to list *v1.Secret: secrets "v4-0-config-user-template-provider-selection" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.678543 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-provider-selection\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.678582 4894 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.678603 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.678490 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678616 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.678630 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678650 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678688 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678779 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678799 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678808 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.678914 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.678980 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.678998 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.679020 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.679120 4894 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.679137 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.679267 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.679358 4894 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-serving-cert": failed to list *v1.Secret: secrets "v4-0-config-system-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.679374 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.679684 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.679697 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.679795 4894 reflector.go:561] object-"openshift-console-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console-operator": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.679814 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.679879 4894 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.679900 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.679945 4894 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.679962 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.680007 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.680027 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680087 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.680211 4894 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.680234 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680287 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680408 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680417 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680530 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680538 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.680624 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.680641 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680669 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680697 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.680781 4894 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-login": failed to list *v1.Secret: secrets "v4-0-config-user-template-login" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.680803 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-login\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.680817 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.680834 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680788 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680874 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v"] Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.680904 4894 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.713115 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.714370 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.714713 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.722209 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.722405 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.680968 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: W1210 18:57:51.681006 4894 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.681008 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.725116 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: E1210 18:57:51.684962 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.737915 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.738882 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.739400 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.739657 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.741150 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.741592 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hqtqc"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.741887 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.742111 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.742237 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.742363 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.742887 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.743160 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.743382 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.743573 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.745476 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.745753 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.745821 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.746248 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.746818 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.747329 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.747908 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k47sq"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.748966 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bl258"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.749409 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.750489 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.751116 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.751239 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vsms7"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.751419 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.752153 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m2s4d"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.753144 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757071 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757099 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757120 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757133 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dwxt5"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757082 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757148 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/85556679-8cbc-4321-a936-22924ac80223-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757173 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-config\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757193 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62khf\" (UniqueName: \"kubernetes.io/projected/85556679-8cbc-4321-a936-22924ac80223-kube-api-access-62khf\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757215 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-config\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757263 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-etcd-serving-ca\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757285 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-dir\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757307 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-encryption-config\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757331 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85556679-8cbc-4321-a936-22924ac80223-config\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757356 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6dhf\" (UniqueName: \"kubernetes.io/projected/c921a149-779b-4ec8-abdd-2b1ed019a0c1-kube-api-access-x6dhf\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757401 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/81c355f5-65db-434b-8af0-068b10da5b92-trusted-ca\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757434 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-oauth-config\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757457 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-image-import-ca\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757481 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/85556679-8cbc-4321-a936-22924ac80223-images\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757506 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757532 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-config\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757581 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtx7g\" (UniqueName: \"kubernetes.io/projected/678d3fbd-5b3d-4e7d-8637-261937d74d01-kube-api-access-mtx7g\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757624 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-service-ca\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757648 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757691 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-etcd-client\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757784 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757805 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/678d3fbd-5b3d-4e7d-8637-261937d74d01-serving-cert\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757828 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757868 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zrtp\" (UniqueName: \"kubernetes.io/projected/d6ce0221-8a80-4e30-bfac-2529f7c2c25d-kube-api-access-6zrtp\") pod \"downloads-7954f5f757-bqdvr\" (UID: \"d6ce0221-8a80-4e30-bfac-2529f7c2c25d\") " pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757889 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757914 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.757976 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758007 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81c355f5-65db-434b-8af0-068b10da5b92-config\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758057 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzwtj\" (UniqueName: \"kubernetes.io/projected/b4094d0b-fce7-4eb9-95c1-174110c39fb2-kube-api-access-xzwtj\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758115 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c355f5-65db-434b-8af0-068b10da5b92-serving-cert\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-serving-cert\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758192 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c921a149-779b-4ec8-abdd-2b1ed019a0c1-node-pullsecrets\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-audit\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758250 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9kck\" (UniqueName: \"kubernetes.io/projected/139baf99-78f8-41f5-8c3c-2d37c42edd47-kube-api-access-w9kck\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.758340 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-serving-cert\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759572 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759614 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-service-ca-bundle\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759643 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759667 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759689 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4094d0b-fce7-4eb9-95c1-174110c39fb2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759710 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-oauth-serving-cert\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759187 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759747 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759778 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759780 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c921a149-779b-4ec8-abdd-2b1ed019a0c1-audit-dir\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759812 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f7kg\" (UniqueName: \"kubernetes.io/projected/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-kube-api-access-5f7kg\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759836 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-policies\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfk4x\" (UniqueName: \"kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759292 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759900 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-dir\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759921 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpqfh\" (UniqueName: \"kubernetes.io/projected/81c355f5-65db-434b-8af0-068b10da5b92-kube-api-access-mpqfh\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759510 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759941 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4094d0b-fce7-4eb9-95c1-174110c39fb2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759955 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759549 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760030 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblkp\" (UniqueName: \"kubernetes.io/projected/02ec8b17-2505-4d2b-a857-c92e5f47b484-kube-api-access-nblkp\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.759579 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760172 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760221 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760296 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760308 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760323 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760403 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.760585 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.761012 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.761403 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.761908 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.762910 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.763401 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hqtqc"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.768697 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.768960 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.769276 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.769406 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.827050 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.830216 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.830487 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.830869 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.836101 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.837028 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bqdvr"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.867409 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.869305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-service-ca\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.870833 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.871961 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898112 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6pnv8"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898135 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898147 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898167 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.877047 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.873952 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-service-ca\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.873980 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898865 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-config\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtx7g\" (UniqueName: \"kubernetes.io/projected/678d3fbd-5b3d-4e7d-8637-261937d74d01-kube-api-access-mtx7g\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898907 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898950 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.898975 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-etcd-client\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899035 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899060 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/678d3fbd-5b3d-4e7d-8637-261937d74d01-serving-cert\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899065 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hm4c5"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899077 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899093 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899107 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899128 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zrtp\" (UniqueName: \"kubernetes.io/projected/d6ce0221-8a80-4e30-bfac-2529f7c2c25d-kube-api-access-6zrtp\") pod \"downloads-7954f5f757-bqdvr\" (UID: \"d6ce0221-8a80-4e30-bfac-2529f7c2c25d\") " pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899144 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899170 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzwtj\" (UniqueName: \"kubernetes.io/projected/b4094d0b-fce7-4eb9-95c1-174110c39fb2-kube-api-access-xzwtj\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899184 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81c355f5-65db-434b-8af0-068b10da5b92-config\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899202 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c355f5-65db-434b-8af0-068b10da5b92-serving-cert\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899236 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c921a149-779b-4ec8-abdd-2b1ed019a0c1-node-pullsecrets\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899253 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-serving-cert\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899268 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899284 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-audit\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899310 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9kck\" (UniqueName: \"kubernetes.io/projected/139baf99-78f8-41f5-8c3c-2d37c42edd47-kube-api-access-w9kck\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899353 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-serving-cert\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899370 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899385 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899409 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899423 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899437 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-service-ca-bundle\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899456 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d0ab405c-e580-4043-a10d-b2dd2f87a906-machine-approver-tls\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899491 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4094d0b-fce7-4eb9-95c1-174110c39fb2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899505 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-oauth-serving-cert\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899519 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-policies\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899533 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfk4x\" (UniqueName: \"kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899548 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c921a149-779b-4ec8-abdd-2b1ed019a0c1-audit-dir\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7kg\" (UniqueName: \"kubernetes.io/projected/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-kube-api-access-5f7kg\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899580 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4094d0b-fce7-4eb9-95c1-174110c39fb2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899601 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hrhlh"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-dir\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899627 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpqfh\" (UniqueName: \"kubernetes.io/projected/81c355f5-65db-434b-8af0-068b10da5b92-kube-api-access-mpqfh\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899645 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblkp\" (UniqueName: \"kubernetes.io/projected/02ec8b17-2505-4d2b-a857-c92e5f47b484-kube-api-access-nblkp\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899667 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/85556679-8cbc-4321-a936-22924ac80223-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899681 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-config\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62khf\" (UniqueName: \"kubernetes.io/projected/85556679-8cbc-4321-a936-22924ac80223-kube-api-access-62khf\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899749 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899771 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0ab405c-e580-4043-a10d-b2dd2f87a906-auth-proxy-config\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-config\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-etcd-serving-ca\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899832 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0ab405c-e580-4043-a10d-b2dd2f87a906-config\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899863 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-dir\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899886 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkh2j\" (UniqueName: \"kubernetes.io/projected/d0ab405c-e580-4043-a10d-b2dd2f87a906-kube-api-access-hkh2j\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899907 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85556679-8cbc-4321-a936-22924ac80223-config\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899926 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-encryption-config\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899963 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/81c355f5-65db-434b-8af0-068b10da5b92-trusted-ca\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.899981 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6dhf\" (UniqueName: \"kubernetes.io/projected/c921a149-779b-4ec8-abdd-2b1ed019a0c1-kube-api-access-x6dhf\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.900013 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/85556679-8cbc-4321-a936-22924ac80223-images\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.900035 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-oauth-config\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.900055 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-image-import-ca\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.900217 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hqkq2"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.900678 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jnckv"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.901024 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.901179 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.901558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-image-import-ca\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.901617 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.903056 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.903098 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.904703 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.905146 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c921a149-779b-4ec8-abdd-2b1ed019a0c1-node-pullsecrets\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.905384 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.905901 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-audit\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.906508 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-config\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.907552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.908274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-oauth-serving-cert\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.908810 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-policies\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.908960 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.908968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c921a149-779b-4ec8-abdd-2b1ed019a0c1-audit-dir\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.909349 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-dir\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.909501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4094d0b-fce7-4eb9-95c1-174110c39fb2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.910042 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-service-ca-bundle\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.910369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/678d3fbd-5b3d-4e7d-8637-261937d74d01-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.910967 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-config\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.911196 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-config\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.911520 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.911558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-etcd-serving-ca\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.911621 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-dir\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.912526 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.913886 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-etcd-client\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.916204 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/85556679-8cbc-4321-a936-22924ac80223-images\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.916690 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81c355f5-65db-434b-8af0-068b10da5b92-config\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.917647 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-serving-cert\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.917914 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-476c9"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.917814 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.918488 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.918500 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85556679-8cbc-4321-a936-22924ac80223-config\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.918740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c921a149-779b-4ec8-abdd-2b1ed019a0c1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.918996 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.919540 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.919652 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/81c355f5-65db-434b-8af0-068b10da5b92-trusted-ca\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.920624 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.924358 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.925024 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.924365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-serving-cert\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.925435 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.925695 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.925761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.926261 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4094d0b-fce7-4eb9-95c1-174110c39fb2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.927678 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/85556679-8cbc-4321-a936-22924ac80223-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.929452 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mk6q8"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.929894 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/678d3fbd-5b3d-4e7d-8637-261937d74d01-serving-cert\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.930164 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.939974 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vpmkx"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.940218 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c921a149-779b-4ec8-abdd-2b1ed019a0c1-encryption-config\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.940886 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.941069 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-oauth-config\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.941452 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.942054 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.942753 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.944041 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.944099 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jkvhm"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.945994 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.946506 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.946693 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.947154 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.947517 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.947823 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jxj6j"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.947925 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.948570 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.949611 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c355f5-65db-434b-8af0-068b10da5b92-serving-cert\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.953944 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.954550 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.955161 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.955714 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.957194 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-z7vst"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.957805 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.958113 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.959789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.963327 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.964533 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.964685 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.965286 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.965630 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.966040 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.966962 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.969296 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wrp99"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.969745 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.970473 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hqkq2"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.972213 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.973390 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hm4c5"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.974453 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.975502 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hrhlh"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.976571 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.978096 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jxj6j"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.979384 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jnckv"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.980637 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.982140 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jkvhm"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.983419 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.983556 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.986589 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-z7vst"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.988168 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-476c9"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.989423 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.990942 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kpk7m"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.991488 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.992023 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.993148 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.994317 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vpmkx"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.995708 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.996768 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.998226 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wrp99"] Dec 10 18:57:51 crc kubenswrapper[4894]: I1210 18:57:51.999897 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv"] Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.000924 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twgpt\" (UniqueName: \"kubernetes.io/projected/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-kube-api-access-twgpt\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.000953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.000979 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzkl6\" (UniqueName: \"kubernetes.io/projected/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-kube-api-access-hzkl6\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001003 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-srv-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001029 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd889abc-66c5-4df6-a208-65d6b64290c2-config\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001050 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-metrics-certs\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001115 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-proxy-tls\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001142 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sw94\" (UniqueName: \"kubernetes.io/projected/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-kube-api-access-4sw94\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/126ef64d-937f-461e-a17a-ad65e8c0d284-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001205 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdswd\" (UniqueName: \"kubernetes.io/projected/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-kube-api-access-fdswd\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001240 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56vzd\" (UniqueName: \"kubernetes.io/projected/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-kube-api-access-56vzd\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001285 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f"] Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001307 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3eb49ab9-5530-4fad-9857-1a74655b56b2-metrics-tls\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-profile-collector-cert\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001639 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001683 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swvm2\" (UniqueName: \"kubernetes.io/projected/26655866-c2c4-4743-9ae9-a1a990701a66-kube-api-access-swvm2\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001738 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d0ab405c-e580-4043-a10d-b2dd2f87a906-machine-approver-tls\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001771 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjfrl\" (UniqueName: \"kubernetes.io/projected/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-kube-api-access-xjfrl\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001835 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001886 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd889abc-66c5-4df6-a208-65d6b64290c2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001926 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3eb49ab9-5530-4fad-9857-1a74655b56b2-config-volume\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-config\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.001990 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-client\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002037 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcfj6\" (UniqueName: \"kubernetes.io/projected/fdbfb107-1ca7-4d48-bed7-382f23014759-kube-api-access-gcfj6\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6gl9\" (UniqueName: \"kubernetes.io/projected/e39c4e75-61aa-4690-bff2-eb4be173b66e-kube-api-access-c6gl9\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002097 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-client-ca\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002131 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-webhook-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttfl\" (UniqueName: \"kubernetes.io/projected/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-kube-api-access-9ttfl\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002187 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxfq2\" (UniqueName: \"kubernetes.io/projected/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-kube-api-access-fxfq2\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002211 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-srv-cert\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002234 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-service-ca\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002258 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5xbv\" (UniqueName: \"kubernetes.io/projected/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-kube-api-access-b5xbv\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqccq\" (UniqueName: \"kubernetes.io/projected/02a170eb-e141-45e7-b26f-3029ce9a5cf2-kube-api-access-jqccq\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002299 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-tmpfs\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002321 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/126ef64d-937f-461e-a17a-ad65e8c0d284-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002363 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002389 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-apiservice-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002439 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-ca\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002461 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-serving-cert\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002500 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/26655866-c2c4-4743-9ae9-a1a990701a66-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002526 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002572 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002594 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdbfb107-1ca7-4d48-bed7-382f23014759-serving-cert\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002636 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/126ef64d-937f-461e-a17a-ad65e8c0d284-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002683 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7h5q\" (UniqueName: \"kubernetes.io/projected/7fe4fbd5-317b-49fd-9c30-82860839a273-kube-api-access-d7h5q\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002705 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-trusted-ca\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002735 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002758 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lcp9\" (UniqueName: \"kubernetes.io/projected/67412df2-b9d8-4713-b76f-6306ea6ed0f4-kube-api-access-8lcp9\") pod \"cluster-samples-operator-665b6dd947-w28xz\" (UID: \"67412df2-b9d8-4713-b76f-6306ea6ed0f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002783 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4cws\" (UniqueName: \"kubernetes.io/projected/78827665-0808-419e-910f-7dce427f3334-kube-api-access-m4cws\") pod \"migrator-59844c95c7-79fm4\" (UID: \"78827665-0808-419e-910f-7dce427f3334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002873 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e39c4e75-61aa-4690-bff2-eb4be173b66e-secret-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002915 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfs9l\" (UniqueName: \"kubernetes.io/projected/600e8028-f30b-423a-b0fc-c5f461d5fea3-kube-api-access-tfs9l\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002952 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-default-certificate\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.002975 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-config\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0ab405c-e580-4043-a10d-b2dd2f87a906-auth-proxy-config\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqvz7\" (UniqueName: \"kubernetes.io/projected/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-kube-api-access-nqvz7\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003120 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003144 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-metrics-tls\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003169 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe4fbd5-317b-49fd-9c30-82860839a273-serving-cert\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003192 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsdct\" (UniqueName: \"kubernetes.io/projected/4141adb8-a6f6-492c-8669-c0742f481d49-kube-api-access-dsdct\") pod \"multus-admission-controller-857f4d67dd-vpmkx\" (UID: \"4141adb8-a6f6-492c-8669-c0742f481d49\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003235 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0ab405c-e580-4043-a10d-b2dd2f87a906-config\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26655866-c2c4-4743-9ae9-a1a990701a66-serving-cert\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-config\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkh2j\" (UniqueName: \"kubernetes.io/projected/d0ab405c-e580-4043-a10d-b2dd2f87a906-kube-api-access-hkh2j\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003500 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/67412df2-b9d8-4713-b76f-6306ea6ed0f4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w28xz\" (UID: \"67412df2-b9d8-4713-b76f-6306ea6ed0f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003530 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02a170eb-e141-45e7-b26f-3029ce9a5cf2-cert\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkdnv\" (UniqueName: \"kubernetes.io/projected/2ef09a89-62d2-43c2-bef7-78e4474f14f4-kube-api-access-tkdnv\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003586 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003653 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4141adb8-a6f6-492c-8669-c0742f481d49-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vpmkx\" (UID: \"4141adb8-a6f6-492c-8669-c0742f481d49\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003688 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd889abc-66c5-4df6-a208-65d6b64290c2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003714 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-service-ca-bundle\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003742 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-stats-auth\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003768 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h79t4\" (UniqueName: \"kubernetes.io/projected/3eb49ab9-5530-4fad-9857-1a74655b56b2-kube-api-access-h79t4\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003769 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0ab405c-e580-4043-a10d-b2dd2f87a906-config\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.003833 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0ab405c-e580-4043-a10d-b2dd2f87a906-auth-proxy-config\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.004493 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.004604 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp"] Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.004758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d0ab405c-e580-4043-a10d-b2dd2f87a906-machine-approver-tls\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.006005 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb"] Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.024201 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.063420 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.068933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtx7g\" (UniqueName: \"kubernetes.io/projected/678d3fbd-5b3d-4e7d-8637-261937d74d01-kube-api-access-mtx7g\") pod \"authentication-operator-69f744f599-k47sq\" (UID: \"678d3fbd-5b3d-4e7d-8637-261937d74d01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.083520 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.103809 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104504 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-proxy-tls\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104537 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sw94\" (UniqueName: \"kubernetes.io/projected/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-kube-api-access-4sw94\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104572 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/126ef64d-937f-461e-a17a-ad65e8c0d284-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdswd\" (UniqueName: \"kubernetes.io/projected/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-kube-api-access-fdswd\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104607 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104624 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56vzd\" (UniqueName: \"kubernetes.io/projected/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-kube-api-access-56vzd\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104643 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3eb49ab9-5530-4fad-9857-1a74655b56b2-metrics-tls\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104664 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-profile-collector-cert\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104689 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104708 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swvm2\" (UniqueName: \"kubernetes.io/projected/26655866-c2c4-4743-9ae9-a1a990701a66-kube-api-access-swvm2\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104727 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjfrl\" (UniqueName: \"kubernetes.io/projected/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-kube-api-access-xjfrl\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104781 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd889abc-66c5-4df6-a208-65d6b64290c2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3eb49ab9-5530-4fad-9857-1a74655b56b2-config-volume\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104821 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-config\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104836 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-client\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104866 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcfj6\" (UniqueName: \"kubernetes.io/projected/fdbfb107-1ca7-4d48-bed7-382f23014759-kube-api-access-gcfj6\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104884 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6gl9\" (UniqueName: \"kubernetes.io/projected/e39c4e75-61aa-4690-bff2-eb4be173b66e-kube-api-access-c6gl9\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104898 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-client-ca\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104949 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-webhook-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104965 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttfl\" (UniqueName: \"kubernetes.io/projected/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-kube-api-access-9ttfl\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxfq2\" (UniqueName: \"kubernetes.io/projected/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-kube-api-access-fxfq2\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.104998 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-srv-cert\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105012 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-service-ca\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5xbv\" (UniqueName: \"kubernetes.io/projected/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-kube-api-access-b5xbv\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105042 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqccq\" (UniqueName: \"kubernetes.io/projected/02a170eb-e141-45e7-b26f-3029ce9a5cf2-kube-api-access-jqccq\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105057 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-tmpfs\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/126ef64d-937f-461e-a17a-ad65e8c0d284-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105094 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105114 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105128 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-apiservice-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105143 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-ca\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105161 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-serving-cert\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/26655866-c2c4-4743-9ae9-a1a990701a66-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105206 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105220 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105234 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdbfb107-1ca7-4d48-bed7-382f23014759-serving-cert\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105263 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/126ef64d-937f-461e-a17a-ad65e8c0d284-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105292 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7h5q\" (UniqueName: \"kubernetes.io/projected/7fe4fbd5-317b-49fd-9c30-82860839a273-kube-api-access-d7h5q\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-trusted-ca\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105459 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lcp9\" (UniqueName: \"kubernetes.io/projected/67412df2-b9d8-4713-b76f-6306ea6ed0f4-kube-api-access-8lcp9\") pod \"cluster-samples-operator-665b6dd947-w28xz\" (UID: \"67412df2-b9d8-4713-b76f-6306ea6ed0f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105483 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4cws\" (UniqueName: \"kubernetes.io/projected/78827665-0808-419e-910f-7dce427f3334-kube-api-access-m4cws\") pod \"migrator-59844c95c7-79fm4\" (UID: \"78827665-0808-419e-910f-7dce427f3334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105515 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e39c4e75-61aa-4690-bff2-eb4be173b66e-secret-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105541 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfs9l\" (UniqueName: \"kubernetes.io/projected/600e8028-f30b-423a-b0fc-c5f461d5fea3-kube-api-access-tfs9l\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105569 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-default-certificate\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105585 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105600 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105615 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-config\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105642 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqvz7\" (UniqueName: \"kubernetes.io/projected/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-kube-api-access-nqvz7\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105658 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105672 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-metrics-tls\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105687 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe4fbd5-317b-49fd-9c30-82860839a273-serving-cert\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105703 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsdct\" (UniqueName: \"kubernetes.io/projected/4141adb8-a6f6-492c-8669-c0742f481d49-kube-api-access-dsdct\") pod \"multus-admission-controller-857f4d67dd-vpmkx\" (UID: \"4141adb8-a6f6-492c-8669-c0742f481d49\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105718 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26655866-c2c4-4743-9ae9-a1a990701a66-serving-cert\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105732 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-config\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105754 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/67412df2-b9d8-4713-b76f-6306ea6ed0f4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w28xz\" (UID: \"67412df2-b9d8-4713-b76f-6306ea6ed0f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105787 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02a170eb-e141-45e7-b26f-3029ce9a5cf2-cert\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkdnv\" (UniqueName: \"kubernetes.io/projected/2ef09a89-62d2-43c2-bef7-78e4474f14f4-kube-api-access-tkdnv\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105822 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105866 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4141adb8-a6f6-492c-8669-c0742f481d49-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vpmkx\" (UID: \"4141adb8-a6f6-492c-8669-c0742f481d49\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105881 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd889abc-66c5-4df6-a208-65d6b64290c2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105896 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-service-ca-bundle\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105910 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105925 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-stats-auth\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105940 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h79t4\" (UniqueName: \"kubernetes.io/projected/3eb49ab9-5530-4fad-9857-1a74655b56b2-kube-api-access-h79t4\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105946 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.105953 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-srv-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twgpt\" (UniqueName: \"kubernetes.io/projected/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-kube-api-access-twgpt\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106051 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106071 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzkl6\" (UniqueName: \"kubernetes.io/projected/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-kube-api-access-hzkl6\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106092 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd889abc-66c5-4df6-a208-65d6b64290c2-config\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106111 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-metrics-certs\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106110 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-tmpfs\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106608 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-client-ca\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.106937 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd889abc-66c5-4df6-a208-65d6b64290c2-config\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.107460 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/26655866-c2c4-4743-9ae9-a1a990701a66-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.108352 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.109437 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-config\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.109655 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3eb49ab9-5530-4fad-9857-1a74655b56b2-config-volume\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.110233 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.111348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.113484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd889abc-66c5-4df6-a208-65d6b64290c2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.114328 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26655866-c2c4-4743-9ae9-a1a990701a66-serving-cert\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.114370 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/67412df2-b9d8-4713-b76f-6306ea6ed0f4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w28xz\" (UID: \"67412df2-b9d8-4713-b76f-6306ea6ed0f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.114414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe4fbd5-317b-49fd-9c30-82860839a273-serving-cert\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.116186 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.124185 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.144792 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.152200 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fdbfb107-1ca7-4d48-bed7-382f23014759-serving-cert\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.163945 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.167996 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3eb49ab9-5530-4fad-9857-1a74655b56b2-metrics-tls\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.183566 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.223513 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.225007 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9kck\" (UniqueName: \"kubernetes.io/projected/139baf99-78f8-41f5-8c3c-2d37c42edd47-kube-api-access-w9kck\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.227863 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-config\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.243622 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.255778 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-client\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.264091 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.266929 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-service-ca\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.277441 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.283683 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.288805 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fdbfb107-1ca7-4d48-bed7-382f23014759-etcd-ca\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.317952 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8e7ecdeb52bf9ac6b8d4af10fa58c057561d8dfecf3adf9d9b202bf326abadff"} Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.319673 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"63f097fe0863ca5efceff1b208658511bad01856d9b7abd1c675d6c1bacc7da2"} Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.320763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"930ebacb13105e4aa6e7aae7afae413143ccaad62a26c327c9acf92587d1f448"} Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.321006 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.321661 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzwtj\" (UniqueName: \"kubernetes.io/projected/b4094d0b-fce7-4eb9-95c1-174110c39fb2-kube-api-access-xzwtj\") pod \"openshift-controller-manager-operator-756b6f6bc6-4r75v\" (UID: \"b4094d0b-fce7-4eb9-95c1-174110c39fb2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.338816 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zrtp\" (UniqueName: \"kubernetes.io/projected/d6ce0221-8a80-4e30-bfac-2529f7c2c25d-kube-api-access-6zrtp\") pod \"downloads-7954f5f757-bqdvr\" (UID: \"d6ce0221-8a80-4e30-bfac-2529f7c2c25d\") " pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.352801 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.364306 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.370327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-metrics-tls\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.414590 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.420977 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-trusted-ca\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.424181 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.479023 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblkp\" (UniqueName: \"kubernetes.io/projected/02ec8b17-2505-4d2b-a857-c92e5f47b484-kube-api-access-nblkp\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.484121 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.504508 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.515813 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k47sq"] Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.541048 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6dhf\" (UniqueName: \"kubernetes.io/projected/c921a149-779b-4ec8-abdd-2b1ed019a0c1-kube-api-access-x6dhf\") pod \"apiserver-76f77b778f-6pnv8\" (UID: \"c921a149-779b-4ec8-abdd-2b1ed019a0c1\") " pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.544328 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.553061 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v"] Dec 10 18:57:52 crc kubenswrapper[4894]: W1210 18:57:52.560799 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4094d0b_fce7_4eb9_95c1_174110c39fb2.slice/crio-5ee6a6a2446e3b3f510ac48ebebee96b6aa7a303945a7ec13b176491ff37e43e WatchSource:0}: Error finding container 5ee6a6a2446e3b3f510ac48ebebee96b6aa7a303945a7ec13b176491ff37e43e: Status 404 returned error can't find the container with id 5ee6a6a2446e3b3f510ac48ebebee96b6aa7a303945a7ec13b176491ff37e43e Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.565621 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.584988 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.604376 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.624115 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.644963 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.664557 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.683673 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.703787 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.724911 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.744784 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.763318 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.769552 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bqdvr"] Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.770686 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/126ef64d-937f-461e-a17a-ad65e8c0d284-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.784060 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.786017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/126ef64d-937f-461e-a17a-ad65e8c0d284-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.786574 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.804940 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.823374 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.833455 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-default-certificate\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.844195 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.852406 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-stats-auth\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.864007 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.874963 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.875061 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.375034888 +0000 UTC m=+156.169882720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.884183 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.888737 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-service-ca-bundle\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.902388 4894 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.902479 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.402453064 +0000 UTC m=+156.197300896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.903572 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.904954 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.904990 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.404980394 +0000 UTC m=+156.199828186 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.905420 4894 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.905479 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.905555 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.4055129 +0000 UTC m=+156.200360802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.905578 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.405569102 +0000 UTC m=+156.200417004 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.908087 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.908164 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.408140273 +0000 UTC m=+156.202988095 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.909268 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.909351 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.40932606 +0000 UTC m=+156.204173882 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.909407 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.909481 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.409460375 +0000 UTC m=+156.204308277 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.910550 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.910647 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.410628642 +0000 UTC m=+156.205476464 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.910669 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.910785 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.410759986 +0000 UTC m=+156.205607818 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.912282 4894 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.912374 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle podName:139baf99-78f8-41f5-8c3c-2d37c42edd47 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.412351286 +0000 UTC m=+156.207199168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle") pod "console-f9d7485db-vsms7" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.913132 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-metrics-certs\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.915108 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.915162 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.415149334 +0000 UTC m=+156.209997246 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.917799 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.918072 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.417944162 +0000 UTC m=+156.212792064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.924200 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.924255 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.924322 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.424303433 +0000 UTC m=+156.219151255 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.925440 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: E1210 18:57:52.929176 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.429108605 +0000 UTC m=+156.223956427 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.942840 4894 request.go:700] Waited for 1.012414509s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.945288 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.986816 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 18:57:52 crc kubenswrapper[4894]: I1210 18:57:52.990338 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62khf\" (UniqueName: \"kubernetes.io/projected/85556679-8cbc-4321-a936-22924ac80223-kube-api-access-62khf\") pod \"machine-api-operator-5694c8668f-bl258\" (UID: \"85556679-8cbc-4321-a936-22924ac80223\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.000762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4141adb8-a6f6-492c-8669-c0742f481d49-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vpmkx\" (UID: \"4141adb8-a6f6-492c-8669-c0742f481d49\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.007353 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.024505 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.044062 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.064838 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.071740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-srv-cert\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.088974 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.101126 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.101840 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e39c4e75-61aa-4690-bff2-eb4be173b66e-secret-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.103996 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.104637 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-profile-collector-cert\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.104709 4894 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.104769 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-proxy-tls podName:ddfa80a5-e9d5-47c5-9514-edda7abb18a6 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.604753206 +0000 UTC m=+156.399600998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-proxy-tls") pod "machine-config-controller-84d6567774-bwnrp" (UID: "ddfa80a5-e9d5-47c5-9514-edda7abb18a6") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.106254 4894 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.106302 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-srv-cert podName:2ef09a89-62d2-43c2-bef7-78e4474f14f4 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.606290304 +0000 UTC m=+156.401138096 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-srv-cert") pod "olm-operator-6b444d44fb-p4wpb" (UID: "2ef09a89-62d2-43c2-bef7-78e4474f14f4") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.106353 4894 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.106390 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-webhook-cert podName:e5f112bf-f37b-4fb8-a610-0bc8026b19b4 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.606381257 +0000 UTC m=+156.401229049 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-webhook-cert") pod "packageserver-d55dfcdfc-cmmx9" (UID: "e5f112bf-f37b-4fb8-a610-0bc8026b19b4") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.107570 4894 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.107617 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-control-plane-machine-set-operator-tls podName:a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.607606596 +0000 UTC m=+156.402454398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-spk6j" (UID: "a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.107679 4894 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.107971 4894 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108021 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-apiservice-cert podName:e5f112bf-f37b-4fb8-a610-0bc8026b19b4 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608009399 +0000 UTC m=+156.402857191 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-apiservice-cert") pod "packageserver-d55dfcdfc-cmmx9" (UID: "e5f112bf-f37b-4fb8-a610-0bc8026b19b4") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108056 4894 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108087 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca podName:600e8028-f30b-423a-b0fc-c5f461d5fea3 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608077341 +0000 UTC m=+156.402925133 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca") pod "marketplace-operator-79b997595-jxj6j" (UID: "600e8028-f30b-423a-b0fc-c5f461d5fea3") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108110 4894 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108138 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-serving-cert podName:06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608129372 +0000 UTC m=+156.402977174 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-serving-cert") pod "service-ca-operator-777779d784-vqxhl" (UID: "06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108162 4894 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108189 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics podName:600e8028-f30b-423a-b0fc-c5f461d5fea3 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608181284 +0000 UTC m=+156.403029076 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics") pod "marketplace-operator-79b997595-jxj6j" (UID: "600e8028-f30b-423a-b0fc-c5f461d5fea3") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108218 4894 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108243 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-config podName:06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608235415 +0000 UTC m=+156.403083207 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-config") pod "service-ca-operator-777779d784-vqxhl" (UID: "06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108264 4894 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108288 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02a170eb-e141-45e7-b26f-3029ce9a5cf2-cert podName:02a170eb-e141-45e7-b26f-3029ce9a5cf2 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608282257 +0000 UTC m=+156.403130049 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/02a170eb-e141-45e7-b26f-3029ce9a5cf2-cert") pod "ingress-canary-z7vst" (UID: "02a170eb-e141-45e7-b26f-3029ce9a5cf2") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.108326 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-config podName:21c5ab50-218c-4b23-a1ef-d7e45a0c668f nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.608318678 +0000 UTC m=+156.403166470 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-config") pod "kube-storage-version-migrator-operator-b67b599dd-f2zrs" (UID: "21c5ab50-218c-4b23-a1ef-d7e45a0c668f") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.109393 4894 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.109438 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-serving-cert podName:21c5ab50-218c-4b23-a1ef-d7e45a0c668f nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.609428273 +0000 UTC m=+156.404276065 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-f2zrs" (UID: "21c5ab50-218c-4b23-a1ef-d7e45a0c668f") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.109532 4894 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.109576 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume podName:e39c4e75-61aa-4690-bff2-eb4be173b66e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.609566178 +0000 UTC m=+156.404413970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume") pod "collect-profiles-29423205-vpp6n" (UID: "e39c4e75-61aa-4690-bff2-eb4be173b66e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.124224 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.144252 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.163971 4894 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.183507 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.197726 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6pnv8"] Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.206654 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.224292 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.244405 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.263884 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.267194 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.302821 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.317565 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.325438 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.345336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bqdvr" event={"ID":"d6ce0221-8a80-4e30-bfac-2529f7c2c25d","Type":"ContainerStarted","Data":"5e0704733ee0a7574c2635a1c65ae5d0b21a52bbc3ca469271d7f2a4551de3ec"} Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.345380 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bqdvr" event={"ID":"d6ce0221-8a80-4e30-bfac-2529f7c2c25d","Type":"ContainerStarted","Data":"1d2c8f426c74b5d1c8eb112aaf921e0e3fe47aef8a4952fd135799bd5ec403ed"} Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.345427 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.345536 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.347399 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-bqdvr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.347448 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bqdvr" podUID="d6ce0221-8a80-4e30-bfac-2529f7c2c25d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.347830 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" event={"ID":"b4094d0b-fce7-4eb9-95c1-174110c39fb2","Type":"ContainerStarted","Data":"59c7ab4f952794edc1f7abb7d7caac1b836b5735a53ee310f500b352a6c4ef81"} Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.347900 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" event={"ID":"b4094d0b-fce7-4eb9-95c1-174110c39fb2","Type":"ContainerStarted","Data":"5ee6a6a2446e3b3f510ac48ebebee96b6aa7a303945a7ec13b176491ff37e43e"} Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.350601 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" event={"ID":"c921a149-779b-4ec8-abdd-2b1ed019a0c1","Type":"ContainerStarted","Data":"fa831c0771316022860a5a6e594e79fed3da814582070f032bd17f53ddca2d38"} Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.353908 4894 projected.go:288] Couldn't get configMap openshift-route-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.360452 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" event={"ID":"678d3fbd-5b3d-4e7d-8637-261937d74d01","Type":"ContainerStarted","Data":"d2b2f2a61ac577c7efa67b3a047ebdd80e484b549efd0e7baa54971ca7a81b4e"} Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.360499 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" event={"ID":"678d3fbd-5b3d-4e7d-8637-261937d74d01","Type":"ContainerStarted","Data":"4325fa555f2aa1a515bbda5e5ee5f54be359b3505cec04ea477dee75db70e4a6"} Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.365892 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.384324 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.417604 4894 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.419767 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.431396 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438274 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438294 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438328 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438368 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438466 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438501 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438566 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438582 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438652 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.438763 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.443567 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.461218 4894 projected.go:288] Couldn't get configMap openshift-console-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.461246 4894 projected.go:194] Error preparing data for projected volume kube-api-access-mpqfh for pod openshift-console-operator/console-operator-58897d9998-dwxt5: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: E1210 18:57:53.461313 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/81c355f5-65db-434b-8af0-068b10da5b92-kube-api-access-mpqfh podName:81c355f5-65db-434b-8af0-068b10da5b92 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:53.961290903 +0000 UTC m=+156.756138695 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mpqfh" (UniqueName: "kubernetes.io/projected/81c355f5-65db-434b-8af0-068b10da5b92-kube-api-access-mpqfh") pod "console-operator-58897d9998-dwxt5" (UID: "81c355f5-65db-434b-8af0-068b10da5b92") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.464586 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.483537 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.504160 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.523866 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.543996 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.546303 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bl258"] Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.563739 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.583579 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.603938 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.623902 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-proxy-tls\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-webhook-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641516 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641537 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-apiservice-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641554 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-serving-cert\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641572 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641661 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641684 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-config\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641713 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641765 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02a170eb-e141-45e7-b26f-3029ce9a5cf2-cert\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641798 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.641819 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-srv-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.642558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-config\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.642807 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.643192 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.643685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.646226 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-proxy-tls\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.646486 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-webhook-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.646641 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02a170eb-e141-45e7-b26f-3029ce9a5cf2-cert\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.647619 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.647786 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-apiservice-cert\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.653746 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-serving-cert\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.654276 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.665172 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.684302 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.698725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2ef09a89-62d2-43c2-bef7-78e4474f14f4-srv-cert\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.707347 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.725730 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.743834 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.765586 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.775243 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.784128 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.793011 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.804422 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.824586 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.844422 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.864161 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.884290 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.904127 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.951943 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.962521 4894 request.go:700] Waited for 1.970783822s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.964159 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 18:57:53 crc kubenswrapper[4894]: I1210 18:57:53.998357 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.024196 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkh2j\" (UniqueName: \"kubernetes.io/projected/d0ab405c-e580-4043-a10d-b2dd2f87a906-kube-api-access-hkh2j\") pod \"machine-approver-56656f9798-8wv22\" (UID: \"d0ab405c-e580-4043-a10d-b2dd2f87a906\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.050201 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56vzd\" (UniqueName: \"kubernetes.io/projected/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-kube-api-access-56vzd\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.059545 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpqfh\" (UniqueName: \"kubernetes.io/projected/81c355f5-65db-434b-8af0-068b10da5b92-kube-api-access-mpqfh\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.068001 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdswd\" (UniqueName: \"kubernetes.io/projected/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-kube-api-access-fdswd\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.090617 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sw94\" (UniqueName: \"kubernetes.io/projected/ddfa80a5-e9d5-47c5-9514-edda7abb18a6-kube-api-access-4sw94\") pod \"machine-config-controller-84d6567774-bwnrp\" (UID: \"ddfa80a5-e9d5-47c5-9514-edda7abb18a6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.109676 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swvm2\" (UniqueName: \"kubernetes.io/projected/26655866-c2c4-4743-9ae9-a1a990701a66-kube-api-access-swvm2\") pod \"openshift-config-operator-7777fb866f-7jmj8\" (UID: \"26655866-c2c4-4743-9ae9-a1a990701a66\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.113630 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.130800 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.133153 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjfrl\" (UniqueName: \"kubernetes.io/projected/e5f112bf-f37b-4fb8-a610-0bc8026b19b4-kube-api-access-xjfrl\") pod \"packageserver-d55dfcdfc-cmmx9\" (UID: \"e5f112bf-f37b-4fb8-a610-0bc8026b19b4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.147529 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twgpt\" (UniqueName: \"kubernetes.io/projected/37a3cb31-fac7-4cc6-aa07-e5db7be4dba4-kube-api-access-twgpt\") pod \"router-default-5444994796-mk6q8\" (UID: \"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4\") " pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:54 crc kubenswrapper[4894]: W1210 18:57:54.148431 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0ab405c_e580_4043_a10d_b2dd2f87a906.slice/crio-d91a4c04bb2c9da4073903dd15d9a35fb79a9325044f35b2790692822d503cbe WatchSource:0}: Error finding container d91a4c04bb2c9da4073903dd15d9a35fb79a9325044f35b2790692822d503cbe: Status 404 returned error can't find the container with id d91a4c04bb2c9da4073903dd15d9a35fb79a9325044f35b2790692822d503cbe Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.181907 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttfl\" (UniqueName: \"kubernetes.io/projected/7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170-kube-api-access-9ttfl\") pod \"catalog-operator-68c6474976-cqk6n\" (UID: \"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.184721 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a0886a7-7bf2-4ba1-94bb-1b740e0064bd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-kvm94\" (UID: \"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.199815 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxfq2\" (UniqueName: \"kubernetes.io/projected/a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4-kube-api-access-fxfq2\") pod \"control-plane-machine-set-operator-78cbb6b69f-spk6j\" (UID: \"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.207282 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.217567 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzkl6\" (UniqueName: \"kubernetes.io/projected/06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc-kube-api-access-hzkl6\") pod \"service-ca-operator-777779d784-vqxhl\" (UID: \"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.225462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.232970 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.246802 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5xbv\" (UniqueName: \"kubernetes.io/projected/21c5ab50-218c-4b23-a1ef-d7e45a0c668f-kube-api-access-b5xbv\") pod \"kube-storage-version-migrator-operator-b67b599dd-f2zrs\" (UID: \"21c5ab50-218c-4b23-a1ef-d7e45a0c668f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.267583 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqccq\" (UniqueName: \"kubernetes.io/projected/02a170eb-e141-45e7-b26f-3029ce9a5cf2-kube-api-access-jqccq\") pod \"ingress-canary-z7vst\" (UID: \"02a170eb-e141-45e7-b26f-3029ce9a5cf2\") " pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.271828 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.276422 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqvz7\" (UniqueName: \"kubernetes.io/projected/0ae28696-fadb-4ac3-8a44-3c1a0163fb90-kube-api-access-nqvz7\") pod \"openshift-apiserver-operator-796bbdcf4f-j6hsz\" (UID: \"0ae28696-fadb-4ac3-8a44-3c1a0163fb90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.278315 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.299605 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/126ef64d-937f-461e-a17a-ad65e8c0d284-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-xt8lh\" (UID: \"126ef64d-937f-461e-a17a-ad65e8c0d284\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.302816 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.315249 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-z7vst" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.317718 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkdnv\" (UniqueName: \"kubernetes.io/projected/2ef09a89-62d2-43c2-bef7-78e4474f14f4-kube-api-access-tkdnv\") pod \"olm-operator-6b444d44fb-p4wpb\" (UID: \"2ef09a89-62d2-43c2-bef7-78e4474f14f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.333362 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.352777 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8"] Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.354113 4894 projected.go:288] Couldn't get configMap openshift-route-controller-manager/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.354130 4894 projected.go:194] Error preparing data for projected volume kube-api-access-sfk4x for pod openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.354185 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:54.854168421 +0000 UTC m=+157.649016213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sfk4x" (UniqueName: "kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.359210 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsdct\" (UniqueName: \"kubernetes.io/projected/4141adb8-a6f6-492c-8669-c0742f481d49-kube-api-access-dsdct\") pod \"multus-admission-controller-857f4d67dd-vpmkx\" (UID: \"4141adb8-a6f6-492c-8669-c0742f481d49\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.364596 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4cws\" (UniqueName: \"kubernetes.io/projected/78827665-0808-419e-910f-7dce427f3334-kube-api-access-m4cws\") pod \"migrator-59844c95c7-79fm4\" (UID: \"78827665-0808-419e-910f-7dce427f3334\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.368436 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mk6q8" event={"ID":"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4","Type":"ContainerStarted","Data":"5272f61d91ac197f513efaea867e35bbd376472b92df615bea773efaaa714bc2"} Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.370957 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" event={"ID":"d0ab405c-e580-4043-a10d-b2dd2f87a906","Type":"ContainerStarted","Data":"d91a4c04bb2c9da4073903dd15d9a35fb79a9325044f35b2790692822d503cbe"} Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.372968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" event={"ID":"85556679-8cbc-4321-a936-22924ac80223","Type":"ContainerStarted","Data":"6b2f0664bce07d8381e6206166faa383648f7e2c9d1449875e8d984b7f65424e"} Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.373114 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" event={"ID":"85556679-8cbc-4321-a936-22924ac80223","Type":"ContainerStarted","Data":"d2c6dfcc9ddcb552cba332ad2b77537a96bb124356fa7fefcceca596c9b6282d"} Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.373129 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" event={"ID":"85556679-8cbc-4321-a936-22924ac80223","Type":"ContainerStarted","Data":"de92dc392694c27e570fdb46b3114e7af211bf8e7dacd573cbc18077cee43395"} Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.374191 4894 generic.go:334] "Generic (PLEG): container finished" podID="c921a149-779b-4ec8-abdd-2b1ed019a0c1" containerID="5701544ba206f47b91f383c5b61cc9ac47e573d70715a50a82b2612990a29212" exitCode=0 Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.374286 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" event={"ID":"c921a149-779b-4ec8-abdd-2b1ed019a0c1","Type":"ContainerDied","Data":"5701544ba206f47b91f383c5b61cc9ac47e573d70715a50a82b2612990a29212"} Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.374976 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-bqdvr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.375020 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bqdvr" podUID="d6ce0221-8a80-4e30-bfac-2529f7c2c25d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 10 18:57:54 crc kubenswrapper[4894]: W1210 18:57:54.375590 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26655866_c2c4_4743_9ae9_a1a990701a66.slice/crio-8e2e000ad346ab411bfa26467461ae49d6a4145d14232edde2cd3de5fd98a3eb WatchSource:0}: Error finding container 8e2e000ad346ab411bfa26467461ae49d6a4145d14232edde2cd3de5fd98a3eb: Status 404 returned error can't find the container with id 8e2e000ad346ab411bfa26467461ae49d6a4145d14232edde2cd3de5fd98a3eb Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.377252 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7h5q\" (UniqueName: \"kubernetes.io/projected/7fe4fbd5-317b-49fd-9c30-82860839a273-kube-api-access-d7h5q\") pod \"controller-manager-879f6c89f-hqtqc\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.401115 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b47e8aa-bfa4-45f8-b46b-54616f3734a4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-f2z8m\" (UID: \"2b47e8aa-bfa4-45f8-b46b-54616f3734a4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.410460 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.417861 4894 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.417891 4894 projected.go:194] Error preparing data for projected volume kube-api-access-5f7kg for pod openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.417954 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-kube-api-access-5f7kg podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:54.917936562 +0000 UTC m=+157.712784354 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5f7kg" (UniqueName: "kubernetes.io/projected/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-kube-api-access-5f7kg") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.426129 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lcp9\" (UniqueName: \"kubernetes.io/projected/67412df2-b9d8-4713-b76f-6306ea6ed0f4-kube-api-access-8lcp9\") pod \"cluster-samples-operator-665b6dd947-w28xz\" (UID: \"67412df2-b9d8-4713-b76f-6306ea6ed0f4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.438256 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438724 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438740 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438787 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.43876753 +0000 UTC m=+158.233615322 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438819 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.438801171 +0000 UTC m=+158.233648963 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438877 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438908 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.438900664 +0000 UTC m=+158.233748536 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438930 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438949 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.438944655 +0000 UTC m=+158.233792447 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438961 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.438978 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.438973026 +0000 UTC m=+158.233820818 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439005 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439030 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439022628 +0000 UTC m=+158.233870520 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439047 4894 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439071 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439063919 +0000 UTC m=+158.233911811 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439096 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439115 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439110591 +0000 UTC m=+158.233958373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439131 4894 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439157 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert podName:02ec8b17-2505-4d2b-a857-c92e5f47b484 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439150463 +0000 UTC m=+158.233998345 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-m2s4d" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439192 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439217 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439208244 +0000 UTC m=+158.234056126 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439231 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439247 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439276 4894 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439287 4894 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439307 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439254 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439247656 +0000 UTC m=+158.234095538 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439341 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439320768 +0000 UTC m=+158.234168560 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439353 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle podName:139baf99-78f8-41f5-8c3c-2d37c42edd47 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439347759 +0000 UTC m=+158.234195551 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle") pod "console-f9d7485db-vsms7" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439364 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert podName:89fbf7f8-3630-4587-95da-e8bf12127a64 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439359279 +0000 UTC m=+158.234207071 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert") pod "route-controller-manager-6576b87f9c-m8r4r" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64") : failed to sync secret cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: E1210 18:57:54.439379 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle podName:2b2c8fcb-b6d9-4d69-be4d-1f14db39761e nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.439372639 +0000 UTC m=+158.234220431 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle") pod "apiserver-7bbb656c7d-t2p24" (UID: "2b2c8fcb-b6d9-4d69-be4d-1f14db39761e") : failed to sync configmap cache: timed out waiting for the condition Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.473074 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.492610 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.494706 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfs9l\" (UniqueName: \"kubernetes.io/projected/600e8028-f30b-423a-b0fc-c5f461d5fea3-kube-api-access-tfs9l\") pod \"marketplace-operator-79b997595-jxj6j\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.500014 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.503469 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h79t4\" (UniqueName: \"kubernetes.io/projected/3eb49ab9-5530-4fad-9857-1a74655b56b2-kube-api-access-h79t4\") pod \"dns-default-hm4c5\" (UID: \"3eb49ab9-5530-4fad-9857-1a74655b56b2\") " pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.508087 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd889abc-66c5-4df6-a208-65d6b64290c2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2fp4b\" (UID: \"cd889abc-66c5-4df6-a208-65d6b64290c2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.509881 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcfj6\" (UniqueName: \"kubernetes.io/projected/fdbfb107-1ca7-4d48-bed7-382f23014759-kube-api-access-gcfj6\") pod \"etcd-operator-b45778765-hqkq2\" (UID: \"fdbfb107-1ca7-4d48-bed7-382f23014759\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.515171 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.519803 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6gl9\" (UniqueName: \"kubernetes.io/projected/e39c4e75-61aa-4690-bff2-eb4be173b66e-kube-api-access-c6gl9\") pod \"collect-profiles-29423205-vpp6n\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.523324 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.543374 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.569411 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.586713 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.591566 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.592173 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.617867 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.628008 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.649548 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.663752 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.694369 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.698399 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.705036 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.724807 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.725237 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.744055 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.745639 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.752050 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.768531 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.769444 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.784153 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.803930 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.824445 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.847095 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n"] Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.848226 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.871046 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.877780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfk4x\" (UniqueName: \"kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.883951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfk4x\" (UniqueName: \"kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.886417 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.889071 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl"] Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.905818 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.923617 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.935410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpqfh\" (UniqueName: \"kubernetes.io/projected/81c355f5-65db-434b-8af0-068b10da5b92-kube-api-access-mpqfh\") pod \"console-operator-58897d9998-dwxt5\" (UID: \"81c355f5-65db-434b-8af0-068b10da5b92\") " pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.946045 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.962574 4894 request.go:700] Waited for 1.646505743s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.967682 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j"] Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.968776 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.976005 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.978372 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs"] Dec 10 18:57:54 crc kubenswrapper[4894]: I1210 18:57:54.979758 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7kg\" (UniqueName: \"kubernetes.io/projected/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-kube-api-access-5f7kg\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.002152 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f7kg\" (UniqueName: \"kubernetes.io/projected/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-kube-api-access-5f7kg\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.007255 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.047818 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.049570 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-z7vst"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083036 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3fce7113-c2a7-4721-ad33-818f861e69ff-signing-key\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083079 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-mountpoint-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083238 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-csi-data-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083261 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnpsn\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-kube-api-access-lnpsn\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083280 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1fa70255-2b2f-4df4-a080-e5fa22f847dd-metrics-tls\") pod \"dns-operator-744455d44c-hrhlh\" (UID: \"1fa70255-2b2f-4df4-a080-e5fa22f847dd\") " pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083306 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d03795b4-dfb1-47b7-9cfe-17f03c460f37-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083479 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e473dc1-ca2e-4f7b-b347-35cce541f595-images\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083579 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwmph\" (UniqueName: \"kubernetes.io/projected/3fce7113-c2a7-4721-ad33-818f861e69ff-kube-api-access-kwmph\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083595 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-plugins-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083616 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-socket-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083637 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9bc3f6-3229-4432-ac8c-ef9395664f67-config\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083764 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9bc3f6-3229-4432-ac8c-ef9395664f67-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083789 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kddwr\" (UniqueName: \"kubernetes.io/projected/2758c256-048d-4ca9-8e76-fc12f1e6301e-kube-api-access-kddwr\") pod \"package-server-manager-789f6589d5-kchjv\" (UID: \"2758c256-048d-4ca9-8e76-fc12f1e6301e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083815 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-certificates\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083829 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-registration-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083949 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d03795b4-dfb1-47b7-9cfe-17f03c460f37-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.083986 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b9bc3f6-3229-4432-ac8c-ef9395664f67-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084002 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-tls\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084044 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-trusted-ca\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e473dc1-ca2e-4f7b-b347-35cce541f595-auth-proxy-config\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084135 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084155 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhgwz\" (UniqueName: \"kubernetes.io/projected/3e473dc1-ca2e-4f7b-b347-35cce541f595-kube-api-access-rhgwz\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084179 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-bound-sa-token\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084196 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2758c256-048d-4ca9-8e76-fc12f1e6301e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kchjv\" (UID: \"2758c256-048d-4ca9-8e76-fc12f1e6301e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084375 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8cpg\" (UniqueName: \"kubernetes.io/projected/1fa70255-2b2f-4df4-a080-e5fa22f847dd-kube-api-access-k8cpg\") pod \"dns-operator-744455d44c-hrhlh\" (UID: \"1fa70255-2b2f-4df4-a080-e5fa22f847dd\") " pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjgrw\" (UniqueName: \"kubernetes.io/projected/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-kube-api-access-tjgrw\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084446 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3fce7113-c2a7-4721-ad33-818f861e69ff-signing-cabundle\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.084575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e473dc1-ca2e-4f7b-b347-35cce541f595-proxy-tls\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.085330 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.585312557 +0000 UTC m=+158.380160429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: W1210 18:57:55.138200 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02a170eb_e141_45e7_b26f_3029ce9a5cf2.slice/crio-164f7ac73ea66b7bf91f980a4061e2454ff623a73a791a166e2120c38ce2ea33 WatchSource:0}: Error finding container 164f7ac73ea66b7bf91f980a4061e2454ff623a73a791a166e2120c38ce2ea33: Status 404 returned error can't find the container with id 164f7ac73ea66b7bf91f980a4061e2454ff623a73a791a166e2120c38ce2ea33 Dec 10 18:57:55 crc kubenswrapper[4894]: W1210 18:57:55.138478 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5f112bf_f37b_4fb8_a610_0bc8026b19b4.slice/crio-b07e35d171132a6838dec6d1e1b289ce4d6d5c299e6241ea9797f2e6b56e58f4 WatchSource:0}: Error finding container b07e35d171132a6838dec6d1e1b289ce4d6d5c299e6241ea9797f2e6b56e58f4: Status 404 returned error can't find the container with id b07e35d171132a6838dec6d1e1b289ce4d6d5c299e6241ea9797f2e6b56e58f4 Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.185427 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.685402025 +0000 UTC m=+158.480249817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185696 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3fce7113-c2a7-4721-ad33-818f861e69ff-signing-cabundle\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185720 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e473dc1-ca2e-4f7b-b347-35cce541f595-proxy-tls\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185807 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3fce7113-c2a7-4721-ad33-818f861e69ff-signing-key\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185823 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-mountpoint-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185867 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-csi-data-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185904 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnpsn\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-kube-api-access-lnpsn\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.185933 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1fa70255-2b2f-4df4-a080-e5fa22f847dd-metrics-tls\") pod \"dns-operator-744455d44c-hrhlh\" (UID: \"1fa70255-2b2f-4df4-a080-e5fa22f847dd\") " pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186011 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d03795b4-dfb1-47b7-9cfe-17f03c460f37-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e473dc1-ca2e-4f7b-b347-35cce541f595-images\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186123 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwmph\" (UniqueName: \"kubernetes.io/projected/3fce7113-c2a7-4721-ad33-818f861e69ff-kube-api-access-kwmph\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186139 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-plugins-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186162 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-socket-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9bc3f6-3229-4432-ac8c-ef9395664f67-config\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186285 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9bc3f6-3229-4432-ac8c-ef9395664f67-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186302 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kddwr\" (UniqueName: \"kubernetes.io/projected/2758c256-048d-4ca9-8e76-fc12f1e6301e-kube-api-access-kddwr\") pod \"package-server-manager-789f6589d5-kchjv\" (UID: \"2758c256-048d-4ca9-8e76-fc12f1e6301e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-certificates\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186477 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-registration-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186520 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d03795b4-dfb1-47b7-9cfe-17f03c460f37-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186580 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b9bc3f6-3229-4432-ac8c-ef9395664f67-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-tls\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186640 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/19b9ebae-f8cd-4f2a-81ef-835eb506408f-node-bootstrap-token\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186667 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/19b9ebae-f8cd-4f2a-81ef-835eb506408f-certs\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186781 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-trusted-ca\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.186878 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e473dc1-ca2e-4f7b-b347-35cce541f595-auth-proxy-config\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187010 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25dp2\" (UniqueName: \"kubernetes.io/projected/19b9ebae-f8cd-4f2a-81ef-835eb506408f-kube-api-access-25dp2\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187280 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhgwz\" (UniqueName: \"kubernetes.io/projected/3e473dc1-ca2e-4f7b-b347-35cce541f595-kube-api-access-rhgwz\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-bound-sa-token\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187394 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2758c256-048d-4ca9-8e76-fc12f1e6301e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kchjv\" (UID: \"2758c256-048d-4ca9-8e76-fc12f1e6301e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187497 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8cpg\" (UniqueName: \"kubernetes.io/projected/1fa70255-2b2f-4df4-a080-e5fa22f847dd-kube-api-access-k8cpg\") pod \"dns-operator-744455d44c-hrhlh\" (UID: \"1fa70255-2b2f-4df4-a080-e5fa22f847dd\") " pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.187522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgrw\" (UniqueName: \"kubernetes.io/projected/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-kube-api-access-tjgrw\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.188479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3e473dc1-ca2e-4f7b-b347-35cce541f595-images\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.189015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3fce7113-c2a7-4721-ad33-818f861e69ff-signing-cabundle\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.191287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d03795b4-dfb1-47b7-9cfe-17f03c460f37-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.193668 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-registration-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.193705 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-plugins-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.193896 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-socket-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.194269 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-mountpoint-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.194359 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-csi-data-dir\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.194645 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e473dc1-ca2e-4f7b-b347-35cce541f595-auth-proxy-config\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.199300 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9bc3f6-3229-4432-ac8c-ef9395664f67-config\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.202255 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.702242106 +0000 UTC m=+158.497089898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.204992 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-certificates\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.205177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-trusted-ca\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.211585 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e473dc1-ca2e-4f7b-b347-35cce541f595-proxy-tls\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.212282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3fce7113-c2a7-4721-ad33-818f861e69ff-signing-key\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.212625 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1fa70255-2b2f-4df4-a080-e5fa22f847dd-metrics-tls\") pod \"dns-operator-744455d44c-hrhlh\" (UID: \"1fa70255-2b2f-4df4-a080-e5fa22f847dd\") " pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.216480 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vpmkx"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.216915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2758c256-048d-4ca9-8e76-fc12f1e6301e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kchjv\" (UID: \"2758c256-048d-4ca9-8e76-fc12f1e6301e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.217104 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-tls\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.223415 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9bc3f6-3229-4432-ac8c-ef9395664f67-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.226626 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.230980 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d03795b4-dfb1-47b7-9cfe-17f03c460f37-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: W1210 18:57:55.244563 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4141adb8_a6f6_492c_8669_c0742f481d49.slice/crio-e659848e00122c7999ede23534e7968c72c7f6c9a75e6f1f37c7b88fb27b2ba4 WatchSource:0}: Error finding container e659848e00122c7999ede23534e7968c72c7f6c9a75e6f1f37c7b88fb27b2ba4: Status 404 returned error can't find the container with id e659848e00122c7999ede23534e7968c72c7f6c9a75e6f1f37c7b88fb27b2ba4 Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.259525 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.261111 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-bound-sa-token\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.262285 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhgwz\" (UniqueName: \"kubernetes.io/projected/3e473dc1-ca2e-4f7b-b347-35cce541f595-kube-api-access-rhgwz\") pod \"machine-config-operator-74547568cd-476c9\" (UID: \"3e473dc1-ca2e-4f7b-b347-35cce541f595\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.263356 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.272426 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hqtqc"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.279436 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8cpg\" (UniqueName: \"kubernetes.io/projected/1fa70255-2b2f-4df4-a080-e5fa22f847dd-kube-api-access-k8cpg\") pod \"dns-operator-744455d44c-hrhlh\" (UID: \"1fa70255-2b2f-4df4-a080-e5fa22f847dd\") " pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.288466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.288830 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25dp2\" (UniqueName: \"kubernetes.io/projected/19b9ebae-f8cd-4f2a-81ef-835eb506408f-kube-api-access-25dp2\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.289104 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/19b9ebae-f8cd-4f2a-81ef-835eb506408f-node-bootstrap-token\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.289123 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/19b9ebae-f8cd-4f2a-81ef-835eb506408f-certs\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.290344 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.790322834 +0000 UTC m=+158.585170626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.295372 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/19b9ebae-f8cd-4f2a-81ef-835eb506408f-certs\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.295831 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/19b9ebae-f8cd-4f2a-81ef-835eb506408f-node-bootstrap-token\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.313131 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgrw\" (UniqueName: \"kubernetes.io/projected/8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b-kube-api-access-tjgrw\") pod \"csi-hostpathplugin-jkvhm\" (UID: \"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b\") " pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.317032 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.318179 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.330368 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnpsn\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-kube-api-access-lnpsn\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.354665 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jxj6j"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.359095 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.363176 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kddwr\" (UniqueName: \"kubernetes.io/projected/2758c256-048d-4ca9-8e76-fc12f1e6301e-kube-api-access-kddwr\") pod \"package-server-manager-789f6589d5-kchjv\" (UID: \"2758c256-048d-4ca9-8e76-fc12f1e6301e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: W1210 18:57:55.368281 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod126ef64d_937f_461e_a17a_ad65e8c0d284.slice/crio-dbd041a91827e7c0f6c10dbd36a2f9f643bf283bd78fa2e52c9a13b02ce7f744 WatchSource:0}: Error finding container dbd041a91827e7c0f6c10dbd36a2f9f643bf283bd78fa2e52c9a13b02ce7f744: Status 404 returned error can't find the container with id dbd041a91827e7c0f6c10dbd36a2f9f643bf283bd78fa2e52c9a13b02ce7f744 Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.385800 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.385926 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwmph\" (UniqueName: \"kubernetes.io/projected/3fce7113-c2a7-4721-ad33-818f861e69ff-kube-api-access-kwmph\") pod \"service-ca-9c57cc56f-wrp99\" (UID: \"3fce7113-c2a7-4721-ad33-818f861e69ff\") " pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.386135 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.389214 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hqkq2"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.393066 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.393398 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b9bc3f6-3229-4432-ac8c-ef9395664f67-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5db7f\" (UID: \"2b9bc3f6-3229-4432-ac8c-ef9395664f67\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.393690 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.893665334 +0000 UTC m=+158.688513126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.410901 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.416696 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" event={"ID":"d0ab405c-e580-4043-a10d-b2dd2f87a906","Type":"ContainerStarted","Data":"286ff4e6af1a9f9ecadfd582d67879d0d2acc0f1f7cc62dac71ba95c078a43b1"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.416944 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" event={"ID":"d0ab405c-e580-4043-a10d-b2dd2f87a906","Type":"ContainerStarted","Data":"58d043011a092f6c807cb40bc5b4eff365715606254e6e483c0ffaa135b4cccf"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.423273 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" event={"ID":"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4","Type":"ContainerStarted","Data":"a24ba3afec6a9f5474c692cf1a14bbcddf3c0491f586c4cd18fa926dbb1b9b22"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.424050 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" event={"ID":"e5f112bf-f37b-4fb8-a610-0bc8026b19b4","Type":"ContainerStarted","Data":"b07e35d171132a6838dec6d1e1b289ce4d6d5c299e6241ea9797f2e6b56e58f4"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.431387 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" event={"ID":"e39c4e75-61aa-4690-bff2-eb4be173b66e","Type":"ContainerStarted","Data":"e7cab4bcc90af69816eb6e943821655e19bb2bc58b1427e61b39bc0e69ff7bab"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.431946 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25dp2\" (UniqueName: \"kubernetes.io/projected/19b9ebae-f8cd-4f2a-81ef-835eb506408f-kube-api-access-25dp2\") pod \"machine-config-server-kpk7m\" (UID: \"19b9ebae-f8cd-4f2a-81ef-835eb506408f\") " pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.438717 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" event={"ID":"0ae28696-fadb-4ac3-8a44-3c1a0163fb90","Type":"ContainerStarted","Data":"c9d8bc894664c9755d9119ee08b18639ebe9beb5623546b1d27821f912d858b0"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.440234 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" event={"ID":"7fe4fbd5-317b-49fd-9c30-82860839a273","Type":"ContainerStarted","Data":"bbd4cec65ea5e1c3bbfd8cccbefae46217d614c6ea8ca21e50b35050aa69b25b"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.441440 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" event={"ID":"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd","Type":"ContainerStarted","Data":"7011d9902f7f34713cf121ff3cd4148a7af76986bf6e962625fa42eae2a6d0c5"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.447511 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" event={"ID":"4141adb8-a6f6-492c-8669-c0742f481d49","Type":"ContainerStarted","Data":"e659848e00122c7999ede23534e7968c72c7f6c9a75e6f1f37c7b88fb27b2ba4"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.454417 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz"] Dec 10 18:57:55 crc kubenswrapper[4894]: W1210 18:57:55.455169 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ef09a89_62d2_43c2_bef7_78e4474f14f4.slice/crio-236ac9b610426281f8c91b5f9cb965f0e4792134dc0f17dc32916aa1629bab80 WatchSource:0}: Error finding container 236ac9b610426281f8c91b5f9cb965f0e4792134dc0f17dc32916aa1629bab80: Status 404 returned error can't find the container with id 236ac9b610426281f8c91b5f9cb965f0e4792134dc0f17dc32916aa1629bab80 Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.455349 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" event={"ID":"78827665-0808-419e-910f-7dce427f3334","Type":"ContainerStarted","Data":"58622361ccaa79c0ee638100854fc03d23f9e641c7e0d721464f7315d9dc9ee0"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.462744 4894 generic.go:334] "Generic (PLEG): container finished" podID="26655866-c2c4-4743-9ae9-a1a990701a66" containerID="b39808932c78284bb9e0775eda188a0266de9b768c17c6b82ecce62072ed653c" exitCode=0 Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.462838 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" event={"ID":"26655866-c2c4-4743-9ae9-a1a990701a66","Type":"ContainerDied","Data":"b39808932c78284bb9e0775eda188a0266de9b768c17c6b82ecce62072ed653c"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.462883 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" event={"ID":"26655866-c2c4-4743-9ae9-a1a990701a66","Type":"ContainerStarted","Data":"8e2e000ad346ab411bfa26467461ae49d6a4145d14232edde2cd3de5fd98a3eb"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.466284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.485695 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mk6q8" event={"ID":"37a3cb31-fac7-4cc6-aa07-e5db7be4dba4","Type":"ContainerStarted","Data":"bb2872564d704c7f826f946e997a0dfa11772c96365df905053b49556a214897"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.488012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" event={"ID":"21c5ab50-218c-4b23-a1ef-d7e45a0c668f","Type":"ContainerStarted","Data":"695500864a04ac9111d9337942d51af9bd79de8e084f26f7c27fdfebe67f13c2"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.489801 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" event={"ID":"ddfa80a5-e9d5-47c5-9514-edda7abb18a6","Type":"ContainerStarted","Data":"757a5aa03e4428a0dd5d1365180af89595e4d9d37710149028c55fd3e22f0847"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.489834 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" event={"ID":"ddfa80a5-e9d5-47c5-9514-edda7abb18a6","Type":"ContainerStarted","Data":"4564d75ed160651a5699d096f0578a0a2aa69edc8e5ad05d5c53bedb718a0a1b"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495145 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495380 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495405 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495466 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495513 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.495532 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496156 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496176 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496197 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496214 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.496249 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.498443 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") pod \"console-f9d7485db-vsms7\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.499215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-audit-policies\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.499293 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:55.999275066 +0000 UTC m=+158.794122858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.499825 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.501403 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.501495 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.501533 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dwxt5"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.502518 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.502734 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.505798 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.508956 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hm4c5"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.523812 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.525108 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-serving-cert\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.525137 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-encryption-config\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.525446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b2c8fcb-b6d9-4d69-be4d-1f14db39761e-etcd-client\") pod \"apiserver-7bbb656c7d-t2p24\" (UID: \"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.525480 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") pod \"route-controller-manager-6576b87f9c-m8r4r\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.525751 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.525880 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.526092 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.527431 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.527696 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m2s4d\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.538912 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.561301 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.563991 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" event={"ID":"c921a149-779b-4ec8-abdd-2b1ed019a0c1","Type":"ContainerStarted","Data":"8562735caf5e2494d64552c0ba66d9b94bd1d3d876398953d10f4d627d361463"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.570943 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kpk7m" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.571746 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" event={"ID":"600e8028-f30b-423a-b0fc-c5f461d5fea3","Type":"ContainerStarted","Data":"17b8dad71997b60bc4bf0f33ee4b9a673c590880d1669749e093a7b7745f590c"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.574154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-z7vst" event={"ID":"02a170eb-e141-45e7-b26f-3029ce9a5cf2","Type":"ContainerStarted","Data":"164f7ac73ea66b7bf91f980a4061e2454ff623a73a791a166e2120c38ce2ea33"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.596571 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.597356 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.598581 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.098568399 +0000 UTC m=+158.893416191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.599218 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" event={"ID":"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170","Type":"ContainerStarted","Data":"6ceff66c69fcd4f7468969d8af91435401ddfb72e0d5a5b784b6a3ac701a6e0c"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.599248 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" event={"ID":"7a4b2e3f-d3a7-48d6-9cb7-8699ccddd170","Type":"ContainerStarted","Data":"18bf67dc86d81c6f9aa187e5546edc7750f32b81236d6b7e6a78c2e7690324e7"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.599526 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.608666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" event={"ID":"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc","Type":"ContainerStarted","Data":"bd5a971de541041bcee0e0f3e3033b16ec3cfe35d4b503951ef97d7b4b7781f6"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.608705 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" event={"ID":"06e2c2b9-5a36-4b54-8546-c5bf7e30fdcc","Type":"ContainerStarted","Data":"da5eb0ff544ecf1029e5a54715059b90583ac3c6325a82919b6efc9b7b4f019b"} Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.609316 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.644193 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.645777 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hrhlh"] Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.702670 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.703021 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.202976982 +0000 UTC m=+158.997824774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.737672 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.806619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.807875 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.307863231 +0000 UTC m=+159.102711023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.816125 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cqk6n" podStartSLOduration=127.816108801 podStartE2EDuration="2m7.816108801s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:55.797878316 +0000 UTC m=+158.592726108" watchObservedRunningTime="2025-12-10 18:57:55.816108801 +0000 UTC m=+158.610956593" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.907607 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-k47sq" podStartSLOduration=127.907591697 podStartE2EDuration="2m7.907591697s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:55.905828012 +0000 UTC m=+158.700675804" watchObservedRunningTime="2025-12-10 18:57:55.907591697 +0000 UTC m=+158.702439489" Dec 10 18:57:55 crc kubenswrapper[4894]: I1210 18:57:55.907857 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:55 crc kubenswrapper[4894]: E1210 18:57:55.908288 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.408271939 +0000 UTC m=+159.203119731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.010224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.010574 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.510557906 +0000 UTC m=+159.305405698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.031942 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-476c9"] Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.033570 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8wv22" podStartSLOduration=128.033555111 podStartE2EDuration="2m8.033555111s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:56.029108481 +0000 UTC m=+158.823956293" watchObservedRunningTime="2025-12-10 18:57:56.033555111 +0000 UTC m=+158.828402903" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.113340 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.114064 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.614044861 +0000 UTC m=+159.408892653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: W1210 18:57:56.200146 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e473dc1_ca2e_4f7b_b347_35cce541f595.slice/crio-cb41f18cb3a41e4123aaa59cdba935d56041c3eb0f2cdfa1a825445c1ecf2813 WatchSource:0}: Error finding container cb41f18cb3a41e4123aaa59cdba935d56041c3eb0f2cdfa1a825445c1ecf2813: Status 404 returned error can't find the container with id cb41f18cb3a41e4123aaa59cdba935d56041c3eb0f2cdfa1a825445c1ecf2813 Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.208256 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.214969 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:56 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:57:56 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:57:56 crc kubenswrapper[4894]: healthz check failed Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.215014 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.216092 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.216389 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.716379879 +0000 UTC m=+159.511227671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.329865 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.330423 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.830406636 +0000 UTC m=+159.625254428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.345817 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jkvhm"] Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.433840 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.434352 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:56.934339505 +0000 UTC m=+159.729187297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.450395 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bqdvr" podStartSLOduration=128.450379761 podStartE2EDuration="2m8.450379761s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:56.449959848 +0000 UTC m=+159.244807660" watchObservedRunningTime="2025-12-10 18:57:56.450379761 +0000 UTC m=+159.245227553" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.544429 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.544778 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.044762018 +0000 UTC m=+159.839609810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.551896 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24"] Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.607244 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mk6q8" podStartSLOduration=128.607227389 podStartE2EDuration="2m8.607227389s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:56.606432203 +0000 UTC m=+159.401279995" watchObservedRunningTime="2025-12-10 18:57:56.607227389 +0000 UTC m=+159.402075181" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.646011 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.646399 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.146386484 +0000 UTC m=+159.941234276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.646957 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hm4c5" event={"ID":"3eb49ab9-5530-4fad-9857-1a74655b56b2","Type":"ContainerStarted","Data":"cbdc557b3909d762bbd829ac2c0c590474c76866c95f32e8c32542954975f973"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.646990 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hm4c5" event={"ID":"3eb49ab9-5530-4fad-9857-1a74655b56b2","Type":"ContainerStarted","Data":"86380c45acd9c1554ba9cb77d2afff3ce2c60b3be56aca8090b1810c558b6132"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.656039 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vqxhl" podStartSLOduration=127.656020498 podStartE2EDuration="2m7.656020498s" podCreationTimestamp="2025-12-10 18:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:56.648670677 +0000 UTC m=+159.443518479" watchObservedRunningTime="2025-12-10 18:57:56.656020498 +0000 UTC m=+159.450868280" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.681102 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" event={"ID":"ddfa80a5-e9d5-47c5-9514-edda7abb18a6","Type":"ContainerStarted","Data":"149b46d9df0db836ba43f99eec0e27d7d9ab07cd2ca30f2eb06a8cac0608260f"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.700032 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" event={"ID":"67412df2-b9d8-4713-b76f-6306ea6ed0f4","Type":"ContainerStarted","Data":"d6956bb1a43eead6b03f8f10acd792d6a38101fca0fa5205144c58fb97947c0f"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.710477 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4r75v" podStartSLOduration=128.710461356 podStartE2EDuration="2m8.710461356s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:56.70899524 +0000 UTC m=+159.503843022" watchObservedRunningTime="2025-12-10 18:57:56.710461356 +0000 UTC m=+159.505309148" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.713065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" event={"ID":"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd","Type":"ContainerStarted","Data":"6e23529f9a89bb8c616ddf2c6db0a12850b5c3101cabee0f6789326eca43ea9b"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.718080 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" event={"ID":"cd889abc-66c5-4df6-a208-65d6b64290c2","Type":"ContainerStarted","Data":"5499fe4560b2d4e4d8d1c5534df66f659630021d77ee6690b23c88fcc3dc56eb"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.730149 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" event={"ID":"e5f112bf-f37b-4fb8-a610-0bc8026b19b4","Type":"ContainerStarted","Data":"0c2e93b913a44bea51381adf58f84e808ea1f62b985d81971f9ca1d63f353d4a"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.740385 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.741275 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" event={"ID":"fdbfb107-1ca7-4d48-bed7-382f23014759","Type":"ContainerStarted","Data":"5fbbc47b3dda249a224eabea6158c5d07028ab0d9045eb117894e136983c1192"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.743596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" event={"ID":"26655866-c2c4-4743-9ae9-a1a990701a66","Type":"ContainerStarted","Data":"f5588b26241601f68c1f33a642c1aff462f52157eb204ceebcc990199b07431d"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.744116 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.748300 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.787979 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.28795893 +0000 UTC m=+160.082806732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.810998 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" event={"ID":"2b47e8aa-bfa4-45f8-b46b-54616f3734a4","Type":"ContainerStarted","Data":"666ef3b02244f8456eed486cd67ea0e8118e7606041f81e651ac1276be2df2a7"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.822002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" event={"ID":"3e473dc1-ca2e-4f7b-b347-35cce541f595","Type":"ContainerStarted","Data":"cb41f18cb3a41e4123aaa59cdba935d56041c3eb0f2cdfa1a825445c1ecf2813"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.867898 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.868501 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.368485051 +0000 UTC m=+160.163332843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.882219 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" event={"ID":"0ae28696-fadb-4ac3-8a44-3c1a0163fb90","Type":"ContainerStarted","Data":"bc357ee610f6c45eca32b2a8265ea11745365e5811c480f1ec5e4f667a9ff746"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.914192 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" event={"ID":"4141adb8-a6f6-492c-8669-c0742f481d49","Type":"ContainerStarted","Data":"a0af409156ac1101b943b4295f2c336d9aea6c547374b788d3ee18aac5a9bdef"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.958115 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" event={"ID":"2ef09a89-62d2-43c2-bef7-78e4474f14f4","Type":"ContainerStarted","Data":"49d0ba068e4dc7d6bd34242021b860b6cdc8ef1998584ebd559ed0bac76d2619"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.958155 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" event={"ID":"2ef09a89-62d2-43c2-bef7-78e4474f14f4","Type":"ContainerStarted","Data":"236ac9b610426281f8c91b5f9cb965f0e4792134dc0f17dc32916aa1629bab80"} Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.960097 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:56 crc kubenswrapper[4894]: I1210 18:57:56.969425 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:56 crc kubenswrapper[4894]: E1210 18:57:56.970717 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.470686495 +0000 UTC m=+160.265534297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.001185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" event={"ID":"81c355f5-65db-434b-8af0-068b10da5b92","Type":"ContainerStarted","Data":"65ab97a56daca5c8ef03d55fcf3dc14c8179924e70abdb5f32bec5127c5a4c46"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.001232 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" event={"ID":"81c355f5-65db-434b-8af0-068b10da5b92","Type":"ContainerStarted","Data":"2379455e98e1e35933106c546baf93e621ca96aed29772f1834adfbba22ff2df"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.033569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" event={"ID":"a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4","Type":"ContainerStarted","Data":"d25370cf6ef21c1c315f71859c96a01bd77568078cb16173cedfb3a1e067175d"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.049267 4894 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p4wpb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.049328 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" podUID="2ef09a89-62d2-43c2-bef7-78e4474f14f4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.050225 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-bl258" podStartSLOduration=129.050214913 podStartE2EDuration="2m9.050214913s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.04945923 +0000 UTC m=+159.844307042" watchObservedRunningTime="2025-12-10 18:57:57.050214913 +0000 UTC m=+159.845062705" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.050553 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" event={"ID":"e39c4e75-61aa-4690-bff2-eb4be173b66e","Type":"ContainerStarted","Data":"1b8571e70614c83cfc2f4179a1b158b448476733f2e469dce6ecac61f18f41fd"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.070492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.071411 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.571399732 +0000 UTC m=+160.366247524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.106141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" event={"ID":"1fa70255-2b2f-4df4-a080-e5fa22f847dd","Type":"ContainerStarted","Data":"c0ee5007a9480dfb04023f235052493cd6fe5c47d564f4f363b01f0d145d6ca0"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.140255 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vr2zg"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.141894 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.146651 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.147384 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" podStartSLOduration=129.147341868 podStartE2EDuration="2m9.147341868s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.146392448 +0000 UTC m=+159.941240250" watchObservedRunningTime="2025-12-10 18:57:57.147341868 +0000 UTC m=+159.942189670" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.172391 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" event={"ID":"c921a149-779b-4ec8-abdd-2b1ed019a0c1","Type":"ContainerStarted","Data":"e1e09a5831fa561bb21731cca632310ccf083ae70db1efb29a083286bfa6c953"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.173419 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.174708 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.67468558 +0000 UTC m=+160.469533422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.186036 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" event={"ID":"78827665-0808-419e-910f-7dce427f3334","Type":"ContainerStarted","Data":"84d9c6b5b31977b52893e3366fe60727d1e888b17e411e6797abd4d1b7b999f1"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.189603 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vr2zg"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.191469 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-j6hsz" podStartSLOduration=129.191446439 podStartE2EDuration="2m9.191446439s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.182205858 +0000 UTC m=+159.977053650" watchObservedRunningTime="2025-12-10 18:57:57.191446439 +0000 UTC m=+159.986294231" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.220137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" event={"ID":"126ef64d-937f-461e-a17a-ad65e8c0d284","Type":"ContainerStarted","Data":"36108939da9054835cc911da72f20f80ab9146948c5ddcf15303c4a6da0d06a4"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.220176 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" event={"ID":"126ef64d-937f-461e-a17a-ad65e8c0d284","Type":"ContainerStarted","Data":"dbd041a91827e7c0f6c10dbd36a2f9f643bf283bd78fa2e52c9a13b02ce7f744"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.220423 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:57 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:57:57 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:57:57 crc kubenswrapper[4894]: healthz check failed Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.220480 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.257420 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kpk7m" event={"ID":"19b9ebae-f8cd-4f2a-81ef-835eb506408f","Type":"ContainerStarted","Data":"71065b48cc2996ab87c9139c269d00323abe1031f04998b3509c6f048705011c"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.265512 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" podStartSLOduration=129.265494675 podStartE2EDuration="2m9.265494675s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.219929758 +0000 UTC m=+160.014777560" watchObservedRunningTime="2025-12-10 18:57:57.265494675 +0000 UTC m=+160.060342467" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.266437 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" podStartSLOduration=129.266431435 podStartE2EDuration="2m9.266431435s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.264829834 +0000 UTC m=+160.059677626" watchObservedRunningTime="2025-12-10 18:57:57.266431435 +0000 UTC m=+160.061279227" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.279515 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-catalog-content\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.279582 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.279652 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb52m\" (UniqueName: \"kubernetes.io/projected/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-kube-api-access-cb52m\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.279742 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-utilities\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.281205 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.7811924 +0000 UTC m=+160.576040192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.291238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" event={"ID":"21c5ab50-218c-4b23-a1ef-d7e45a0c668f","Type":"ContainerStarted","Data":"ddbe0227e85c3995d8fbd64cf6bae39e45a823afc5cfcab229367fa2c2cee49f"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.310159 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bwnrp" podStartSLOduration=129.310146294 podStartE2EDuration="2m9.310146294s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.309520244 +0000 UTC m=+160.104368036" watchObservedRunningTime="2025-12-10 18:57:57.310146294 +0000 UTC m=+160.104994086" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.314382 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zdvpp"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.316012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" event={"ID":"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b","Type":"ContainerStarted","Data":"79fb7485e89271549e3de53320328cc42f2a9092b2568e776e35672712aa0ec2"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.323012 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.328445 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zdvpp"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.334129 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-z7vst" event={"ID":"02a170eb-e141-45e7-b26f-3029ce9a5cf2","Type":"ContainerStarted","Data":"baa29b14a2022bb387071d74fa4b9c8f79c2ce42c2d0dc689a283fada2b97faa"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.344372 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.349491 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" event={"ID":"7fe4fbd5-317b-49fd-9c30-82860839a273","Type":"ContainerStarted","Data":"89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.350828 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.357476 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" podStartSLOduration=129.357455246 podStartE2EDuration="2m9.357455246s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.34457624 +0000 UTC m=+160.139424032" watchObservedRunningTime="2025-12-10 18:57:57.357455246 +0000 UTC m=+160.152303038" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.383674 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.385094 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb52m\" (UniqueName: \"kubernetes.io/projected/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-kube-api-access-cb52m\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.385165 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-utilities\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.385215 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-catalog-content\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.386467 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.886449201 +0000 UTC m=+160.681296993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.390487 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-utilities\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.399489 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.402017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-catalog-content\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.432569 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" podStartSLOduration=129.432551415 podStartE2EDuration="2m9.432551415s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.413979099 +0000 UTC m=+160.208826901" watchObservedRunningTime="2025-12-10 18:57:57.432551415 +0000 UTC m=+160.227399207" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.433785 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jxj6j container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.434306 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.435127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" event={"ID":"600e8028-f30b-423a-b0fc-c5f461d5fea3","Type":"ContainerStarted","Data":"838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c"} Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.435664 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.446581 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.497438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.497497 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzbw7\" (UniqueName: \"kubernetes.io/projected/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-kube-api-access-mzbw7\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.497663 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-utilities\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.497728 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-catalog-content\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.498075 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:57.998063842 +0000 UTC m=+160.792911634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.509871 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-spk6j" podStartSLOduration=129.509831853 podStartE2EDuration="2m9.509831853s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.496351938 +0000 UTC m=+160.291199750" watchObservedRunningTime="2025-12-10 18:57:57.509831853 +0000 UTC m=+160.304679645" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.518554 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb52m\" (UniqueName: \"kubernetes.io/projected/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-kube-api-access-cb52m\") pod \"community-operators-vr2zg\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.527773 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-f2zrs" podStartSLOduration=129.527756008 podStartE2EDuration="2m9.527756008s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.527039786 +0000 UTC m=+160.321887588" watchObservedRunningTime="2025-12-10 18:57:57.527756008 +0000 UTC m=+160.322603820" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.538492 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.607774 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m2s4d"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.625336 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cmmx9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.625362 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m7tz9"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.626214 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m7tz9"] Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.616514 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.116500278 +0000 UTC m=+160.911348070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.616261 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.628076 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-utilities\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.628179 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-catalog-content\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.628282 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.628311 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzbw7\" (UniqueName: \"kubernetes.io/projected/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-kube-api-access-mzbw7\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.621617 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-z7vst" podStartSLOduration=6.621595188 podStartE2EDuration="6.621595188s" podCreationTimestamp="2025-12-10 18:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.606814313 +0000 UTC m=+160.401662115" watchObservedRunningTime="2025-12-10 18:57:57.621595188 +0000 UTC m=+160.416442980" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.632127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-utilities\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.633169 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.133152584 +0000 UTC m=+160.928000376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.636299 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wrp99"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.639312 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.655041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-catalog-content\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.660836 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-xt8lh" podStartSLOduration=129.660814107 podStartE2EDuration="2m9.660814107s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.642888261 +0000 UTC m=+160.437736053" watchObservedRunningTime="2025-12-10 18:57:57.660814107 +0000 UTC m=+160.455661899" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.695918 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzbw7\" (UniqueName: \"kubernetes.io/projected/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-kube-api-access-mzbw7\") pod \"certified-operators-zdvpp\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.702866 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" podStartSLOduration=129.70279194 podStartE2EDuration="2m9.70279194s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.680497468 +0000 UTC m=+160.475345270" watchObservedRunningTime="2025-12-10 18:57:57.70279194 +0000 UTC m=+160.497639742" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.715096 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dkr5p"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.717539 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.730954 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dkr5p"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.735809 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.736069 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-utilities\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.736122 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqxz4\" (UniqueName: \"kubernetes.io/projected/3a0cf23c-9387-4ed5-8f07-95b54c449786-kube-api-access-bqxz4\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.736171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-utilities\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.736209 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.236186624 +0000 UTC m=+161.031034406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.736253 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-catalog-content\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.736283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljt58\" (UniqueName: \"kubernetes.io/projected/dada86bd-f99d-4566-98eb-33194725df47-kube-api-access-ljt58\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.736303 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-catalog-content\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.786644 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.786943 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.800737 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" podStartSLOduration=129.800720779 podStartE2EDuration="2m9.800720779s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.782822966 +0000 UTC m=+160.577670768" watchObservedRunningTime="2025-12-10 18:57:57.800720779 +0000 UTC m=+160.595568571" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.802023 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vsms7"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838557 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-utilities\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838634 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-catalog-content\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838648 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljt58\" (UniqueName: \"kubernetes.io/projected/dada86bd-f99d-4566-98eb-33194725df47-kube-api-access-ljt58\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838667 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-catalog-content\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838687 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-utilities\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.838728 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqxz4\" (UniqueName: \"kubernetes.io/projected/3a0cf23c-9387-4ed5-8f07-95b54c449786-kube-api-access-bqxz4\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.839210 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.339200194 +0000 UTC m=+161.134047986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.839529 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-utilities\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.839743 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-catalog-content\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.840184 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-catalog-content\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.840416 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-utilities\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.850369 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.851210 4894 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6pnv8 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]log ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]etcd ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/generic-apiserver-start-informers ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/max-in-flight-filter ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 10 18:57:57 crc kubenswrapper[4894]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/project.openshift.io-projectcache ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-startinformers ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 10 18:57:57 crc kubenswrapper[4894]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 10 18:57:57 crc kubenswrapper[4894]: livez check failed Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.851249 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" podUID="c921a149-779b-4ec8-abdd-2b1ed019a0c1" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.870863 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f"] Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.903769 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" podStartSLOduration=129.90374937 podStartE2EDuration="2m9.90374937s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:57.900004752 +0000 UTC m=+160.694852564" watchObservedRunningTime="2025-12-10 18:57:57.90374937 +0000 UTC m=+160.698597162" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.915637 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.928880 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljt58\" (UniqueName: \"kubernetes.io/projected/dada86bd-f99d-4566-98eb-33194725df47-kube-api-access-ljt58\") pod \"certified-operators-dkr5p\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.929589 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqxz4\" (UniqueName: \"kubernetes.io/projected/3a0cf23c-9387-4ed5-8f07-95b54c449786-kube-api-access-bqxz4\") pod \"community-operators-m7tz9\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:57 crc kubenswrapper[4894]: I1210 18:57:57.944490 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:57 crc kubenswrapper[4894]: E1210 18:57:57.944754 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.444741243 +0000 UTC m=+161.239589035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.047205 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.047926 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.547886507 +0000 UTC m=+161.342734299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.048109 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.083764 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.159091 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.159716 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.659692845 +0000 UTC m=+161.454540637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.235828 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vr2zg"] Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.253695 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:58 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:57:58 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:57:58 crc kubenswrapper[4894]: healthz check failed Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.253957 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.260707 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.261212 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.761199946 +0000 UTC m=+161.556047738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.361715 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.362093 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.862067678 +0000 UTC m=+161.656915470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.464308 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kpk7m" event={"ID":"19b9ebae-f8cd-4f2a-81ef-835eb506408f","Type":"ContainerStarted","Data":"081a98efa20c747d764142c302e3ebbe9b62933de3497da20b58256039de31a7"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.465454 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.465780 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:58.96576951 +0000 UTC m=+161.760617302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.467690 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" event={"ID":"2b47e8aa-bfa4-45f8-b46b-54616f3734a4","Type":"ContainerStarted","Data":"ce7f0d66a20ef37b45c0f2e78f7ada5effd714b8ab0fc6cef980ffb7d9a97c0e"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.485115 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" event={"ID":"3e473dc1-ca2e-4f7b-b347-35cce541f595","Type":"ContainerStarted","Data":"4fe06ac60a2871bccc9e685fc181b0ee79824c370b5295a82e50599d8ae4259e"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.485162 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" event={"ID":"3e473dc1-ca2e-4f7b-b347-35cce541f595","Type":"ContainerStarted","Data":"d99e7e3b90fd37184d115894fb70aa897321963d81070d0631e167e3f537a5bd"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.506447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vsms7" event={"ID":"139baf99-78f8-41f5-8c3c-2d37c42edd47","Type":"ContainerStarted","Data":"261ce7371bbe067233bc7f45cc17e5e9464c73055de09905b019ba2f16088d27"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.539152 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" event={"ID":"67412df2-b9d8-4713-b76f-6306ea6ed0f4","Type":"ContainerStarted","Data":"9932e2aac925956925c29712fdc8b1acc5e7f4af32e9273bfcda42df9763456f"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.539428 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" event={"ID":"67412df2-b9d8-4713-b76f-6306ea6ed0f4","Type":"ContainerStarted","Data":"14b735409b340a670b6bb0d727bb531f9ad99c992a2df0e4f68eaf816fff30ec"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.589865 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.591565 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.091546358 +0000 UTC m=+161.886394150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.591653 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.598554 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.098495807 +0000 UTC m=+161.893343599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.605724 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" event={"ID":"1a0886a7-7bf2-4ba1-94bb-1b740e0064bd","Type":"ContainerStarted","Data":"e10fed754ab907f6a924d0c2cce9a9052af6241b2a917053f2b7260b0fd4bab7"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.633059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" event={"ID":"89fbf7f8-3630-4587-95da-e8bf12127a64","Type":"ContainerStarted","Data":"b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.633103 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" event={"ID":"89fbf7f8-3630-4587-95da-e8bf12127a64","Type":"ContainerStarted","Data":"66a16666c09f9120951338b28ff86c87742730c7f16b3c4d67e31dae212767b7"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.633784 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.660030 4894 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m8r4r container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.660095 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" podUID="89fbf7f8-3630-4587-95da-e8bf12127a64" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.688086 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" event={"ID":"fdbfb107-1ca7-4d48-bed7-382f23014759","Type":"ContainerStarted","Data":"df5446a2d2d197c291fccd9ee58f4cbe7a9b3ee178885982d21b510c77538ce2"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.692481 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.693384 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.1933703 +0000 UTC m=+161.988218092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.701104 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" event={"ID":"02ec8b17-2505-4d2b-a857-c92e5f47b484","Type":"ContainerStarted","Data":"948289cdc42e6c91dd2a9cafe6322554abaad57551817f6b7c9067a8cf4e0db8"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.724537 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" event={"ID":"2b9bc3f6-3229-4432-ac8c-ef9395664f67","Type":"ContainerStarted","Data":"a6246defc2e0b01914c969bd1c75b8b370c37410edc395e194ccc5599da72c69"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.771962 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" event={"ID":"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b","Type":"ContainerStarted","Data":"7d309551f172aca4d8cb9538917e96596eeb56a1b6727829d1c3bfd982912717"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.800387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.807354 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.307341796 +0000 UTC m=+162.102189588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.843271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" event={"ID":"cd889abc-66c5-4df6-a208-65d6b64290c2","Type":"ContainerStarted","Data":"b93b9e38acc22d5374ac5183d3f608250ae69e700727a96ea7a9cf864107ed57"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.854782 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" event={"ID":"2758c256-048d-4ca9-8e76-fc12f1e6301e","Type":"ContainerStarted","Data":"d6c9a76ea9ea3cf385b0e6a5844ab6cee186a767ffc9649c4702d549e332ffd7"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.871538 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kpk7m" podStartSLOduration=7.87151541 podStartE2EDuration="7.87151541s" podCreationTimestamp="2025-12-10 18:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:58.863112435 +0000 UTC m=+161.657960237" watchObservedRunningTime="2025-12-10 18:57:58.87151541 +0000 UTC m=+161.666363202" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.883044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" event={"ID":"4141adb8-a6f6-492c-8669-c0742f481d49","Type":"ContainerStarted","Data":"ea4f6fc04b6e3e932e0803d0eea2cb811cc84f5a4b088149c775d66f9ff35bb0"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.904676 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.905383 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w28xz" podStartSLOduration=130.905361408 podStartE2EDuration="2m10.905361408s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:58.902498038 +0000 UTC m=+161.697345840" watchObservedRunningTime="2025-12-10 18:57:58.905361408 +0000 UTC m=+161.700209200" Dec 10 18:57:58 crc kubenswrapper[4894]: E1210 18:57:58.907919 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.407901838 +0000 UTC m=+162.202749620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.969351 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f2z8m" podStartSLOduration=130.969338976 podStartE2EDuration="2m10.969338976s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:58.966772855 +0000 UTC m=+161.761620667" watchObservedRunningTime="2025-12-10 18:57:58.969338976 +0000 UTC m=+161.764186768" Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.982000 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m7tz9"] Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.983835 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" event={"ID":"1fa70255-2b2f-4df4-a080-e5fa22f847dd","Type":"ContainerStarted","Data":"398e6e35a2ed3646e037b2d47fa3c281ee1ea14e80e09f02c3d7a264ab32b36e"} Dec 10 18:57:58 crc kubenswrapper[4894]: I1210 18:57:58.985659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" event={"ID":"78827665-0808-419e-910f-7dce427f3334","Type":"ContainerStarted","Data":"816318d57aceb4ad7e40abef518042cdfea52e1795592446e6a61dcff1f32e0d"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.008566 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.008870 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.508833792 +0000 UTC m=+162.303681584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.023307 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hm4c5" event={"ID":"3eb49ab9-5530-4fad-9857-1a74655b56b2","Type":"ContainerStarted","Data":"8fb83ea2b4513b5e9c6125e7a4dcc8cdf190b92b8a50a4c0e7541f00fe44ccc7"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.023921 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hm4c5" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.024982 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-kvm94" podStartSLOduration=131.024972912 podStartE2EDuration="2m11.024972912s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.023687351 +0000 UTC m=+161.818535143" watchObservedRunningTime="2025-12-10 18:57:59.024972912 +0000 UTC m=+161.819820704" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.047232 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b2c8fcb-b6d9-4d69-be4d-1f14db39761e" containerID="6b4068da62e8a20266d4028b0065f2d8cefc6e13382eb38bc6185d9f07cdac5b" exitCode=0 Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.047338 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" event={"ID":"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e","Type":"ContainerDied","Data":"6b4068da62e8a20266d4028b0065f2d8cefc6e13382eb38bc6185d9f07cdac5b"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.047363 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" event={"ID":"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e","Type":"ContainerStarted","Data":"5d3cf29fce9916715bd2369daa6d4937c828b2915c06bac7048eb5dcdc063627"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.054222 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" event={"ID":"3fce7113-c2a7-4721-ad33-818f861e69ff","Type":"ContainerStarted","Data":"0cd3e470a7b8c7e29f3d02a53a288ec6600241f931fb9cb5e7600d8fff7d0314"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.054259 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" event={"ID":"3fce7113-c2a7-4721-ad33-818f861e69ff","Type":"ContainerStarted","Data":"3ece41f02512db013db49f732b6f8d402663ff48d5ee136c4bfec1c5b8234609"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.062181 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hqkq2" podStartSLOduration=131.062163704 podStartE2EDuration="2m11.062163704s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.053801841 +0000 UTC m=+161.848649633" watchObservedRunningTime="2025-12-10 18:57:59.062163704 +0000 UTC m=+161.857011496" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.063259 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2zg" event={"ID":"2a2f64bf-a1f8-4ad3-9dd3-970afc101879","Type":"ContainerStarted","Data":"24376163083bc8fd9c2583bf90caf3f99078d21c2403c9fb9d29d7fee5a8da4c"} Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.065742 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jxj6j container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.065781 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.066144 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.071757 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p4wpb" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.081440 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-476c9" podStartSLOduration=131.081420802 podStartE2EDuration="2m11.081420802s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.077646513 +0000 UTC m=+161.872494305" watchObservedRunningTime="2025-12-10 18:57:59.081420802 +0000 UTC m=+161.876268594" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.103479 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9mb95"] Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.104543 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.111572 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" podStartSLOduration=130.111558903 podStartE2EDuration="2m10.111558903s" podCreationTimestamp="2025-12-10 18:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.10892131 +0000 UTC m=+161.903769112" watchObservedRunningTime="2025-12-10 18:57:59.111558903 +0000 UTC m=+161.906406695" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.112057 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.113190 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.115357 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.615340132 +0000 UTC m=+162.410187924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.123034 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7jmj8" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.139236 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mb95"] Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.179043 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2fp4b" podStartSLOduration=131.179006681 podStartE2EDuration="2m11.179006681s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.177454812 +0000 UTC m=+161.972302614" watchObservedRunningTime="2025-12-10 18:57:59.179006681 +0000 UTC m=+161.973854463" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.204326 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vpmkx" podStartSLOduration=131.204313179 podStartE2EDuration="2m11.204313179s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.203242505 +0000 UTC m=+161.998090307" watchObservedRunningTime="2025-12-10 18:57:59.204313179 +0000 UTC m=+161.999160971" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.216508 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-utilities\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.216774 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-catalog-content\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.216960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.217131 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5867z\" (UniqueName: \"kubernetes.io/projected/91c2b024-f9b4-4d85-bc5d-ad463ae51871-kube-api-access-5867z\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.220433 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.720397096 +0000 UTC m=+162.515244948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.228128 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:59 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:57:59 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:57:59 crc kubenswrapper[4894]: healthz check failed Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.228162 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.234528 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hm4c5" podStartSLOduration=8.234513472 podStartE2EDuration="8.234513472s" podCreationTimestamp="2025-12-10 18:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.232397935 +0000 UTC m=+162.027245727" watchObservedRunningTime="2025-12-10 18:57:59.234513472 +0000 UTC m=+162.029361264" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.237657 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zdvpp"] Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.274712 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dwxt5" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.275465 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-79fm4" podStartSLOduration=131.275454504 podStartE2EDuration="2m11.275454504s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.275137623 +0000 UTC m=+162.069985415" watchObservedRunningTime="2025-12-10 18:57:59.275454504 +0000 UTC m=+162.070302296" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.328982 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.329348 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5867z\" (UniqueName: \"kubernetes.io/projected/91c2b024-f9b4-4d85-bc5d-ad463ae51871-kube-api-access-5867z\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.329402 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-utilities\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.329455 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-catalog-content\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.330796 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.830768218 +0000 UTC m=+162.625616010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.335950 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-catalog-content\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.341179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-utilities\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: W1210 18:57:59.366540 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddada86bd_f99d_4566_98eb_33194725df47.slice/crio-d3070f617cf8cef40c9c2b50547f3db6ad3264ad1acb6432eeceb016b41fcd82 WatchSource:0}: Error finding container d3070f617cf8cef40c9c2b50547f3db6ad3264ad1acb6432eeceb016b41fcd82: Status 404 returned error can't find the container with id d3070f617cf8cef40c9c2b50547f3db6ad3264ad1acb6432eeceb016b41fcd82 Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.367545 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dkr5p"] Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.368464 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" podStartSLOduration=131.368442197 podStartE2EDuration="2m11.368442197s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.367000462 +0000 UTC m=+162.161848254" watchObservedRunningTime="2025-12-10 18:57:59.368442197 +0000 UTC m=+162.163289989" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.432539 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.432831 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:59.932819698 +0000 UTC m=+162.727667490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.433238 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5867z\" (UniqueName: \"kubernetes.io/projected/91c2b024-f9b4-4d85-bc5d-ad463ae51871-kube-api-access-5867z\") pod \"redhat-marketplace-9mb95\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.458839 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wrp99" podStartSLOduration=130.458812918 podStartE2EDuration="2m10.458812918s" podCreationTimestamp="2025-12-10 18:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:59.425248379 +0000 UTC m=+162.220096171" watchObservedRunningTime="2025-12-10 18:57:59.458812918 +0000 UTC m=+162.253660710" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.536058 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.536427 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.036413106 +0000 UTC m=+162.831260898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.561293 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kt5vd"] Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.570243 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kt5vd"] Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.570348 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.637482 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.638028 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.138016971 +0000 UTC m=+162.932864763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.727686 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.739012 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.739240 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.239198873 +0000 UTC m=+163.034046665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.739333 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-utilities\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.739455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-catalog-content\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.739524 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.739594 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvwcp\" (UniqueName: \"kubernetes.io/projected/f8cb70e6-c927-4dea-8184-ae0a5343ddec-kube-api-access-jvwcp\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.739918 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.239910985 +0000 UTC m=+163.034758777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.841528 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.841744 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-catalog-content\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.841802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvwcp\" (UniqueName: \"kubernetes.io/projected/f8cb70e6-c927-4dea-8184-ae0a5343ddec-kube-api-access-jvwcp\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.841832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-utilities\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.842176 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.342163701 +0000 UTC m=+163.137011483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.842209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-utilities\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.842321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-catalog-content\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.915015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvwcp\" (UniqueName: \"kubernetes.io/projected/f8cb70e6-c927-4dea-8184-ae0a5343ddec-kube-api-access-jvwcp\") pod \"redhat-marketplace-kt5vd\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:57:59 crc kubenswrapper[4894]: I1210 18:57:59.943722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:57:59 crc kubenswrapper[4894]: E1210 18:57:59.944366 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.444354745 +0000 UTC m=+163.239202537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.045154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.045499 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.545482636 +0000 UTC m=+163.340330428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.073789 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" event={"ID":"2758c256-048d-4ca9-8e76-fc12f1e6301e","Type":"ContainerStarted","Data":"affa2694f6158afefe99b5300803339566153c0729e834bf70e8ac79afc970c8"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.073840 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" event={"ID":"2758c256-048d-4ca9-8e76-fc12f1e6301e","Type":"ContainerStarted","Data":"9b004f94cdf9c77fbb1e6fc52f7b531a808a81b84e130d9f22796a2dfd358f2a"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.073903 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.098208 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" event={"ID":"2b2c8fcb-b6d9-4d69-be4d-1f14db39761e","Type":"ContainerStarted","Data":"c77a840d708dd17d2f6b5cabcfac42f9f1ea399f53a574b071e0cd9a938ed20c"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.121037 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" podStartSLOduration=132.121021159 podStartE2EDuration="2m12.121021159s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:00.120037758 +0000 UTC m=+162.914885540" watchObservedRunningTime="2025-12-10 18:58:00.121021159 +0000 UTC m=+162.915868951" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.121621 4894 generic.go:334] "Generic (PLEG): container finished" podID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerID="424d110b73b06af3a13f063dac8cd6004c0333e9a1085e49be0e72c0d5a689a5" exitCode=0 Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.121678 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7tz9" event={"ID":"3a0cf23c-9387-4ed5-8f07-95b54c449786","Type":"ContainerDied","Data":"424d110b73b06af3a13f063dac8cd6004c0333e9a1085e49be0e72c0d5a689a5"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.121702 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7tz9" event={"ID":"3a0cf23c-9387-4ed5-8f07-95b54c449786","Type":"ContainerStarted","Data":"ff0986d43253fe4b7a0574e3b11f952e174c2ad751e4f2d5f0f4db4d52c0da1b"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.128443 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.147353 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vsms7" event={"ID":"139baf99-78f8-41f5-8c3c-2d37c42edd47","Type":"ContainerStarted","Data":"5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.152593 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.152894 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.652882404 +0000 UTC m=+163.447730196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.169124 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" podStartSLOduration=132.169106866 podStartE2EDuration="2m12.169106866s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:00.167403322 +0000 UTC m=+162.962251114" watchObservedRunningTime="2025-12-10 18:58:00.169106866 +0000 UTC m=+162.963954658" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.188120 4894 generic.go:334] "Generic (PLEG): container finished" podID="dada86bd-f99d-4566-98eb-33194725df47" containerID="680b030bfd1523df05bb4649b20e1b41efc2252370e3ae33eb4ea034972c813b" exitCode=0 Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.188212 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerDied","Data":"680b030bfd1523df05bb4649b20e1b41efc2252370e3ae33eb4ea034972c813b"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.188238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerStarted","Data":"d3070f617cf8cef40c9c2b50547f3db6ad3264ad1acb6432eeceb016b41fcd82"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.193023 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vsms7" podStartSLOduration=132.19300891 podStartE2EDuration="2m12.19300891s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:00.192813763 +0000 UTC m=+162.987661555" watchObservedRunningTime="2025-12-10 18:58:00.19300891 +0000 UTC m=+162.987856702" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.206720 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.214888 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerID="0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5" exitCode=0 Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.214959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2zg" event={"ID":"2a2f64bf-a1f8-4ad3-9dd3-970afc101879","Type":"ContainerDied","Data":"0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.217977 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:00 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:00 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:00 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.218013 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.231057 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mb95"] Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.237767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" event={"ID":"02ec8b17-2505-4d2b-a857-c92e5f47b484","Type":"ContainerStarted","Data":"86f6aebe6299fbe188e18f87065f8123e3c5f839ec5c675a216585181059b5b4"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.237813 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.253147 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.254325 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.754310604 +0000 UTC m=+163.549158396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.254741 4894 generic.go:334] "Generic (PLEG): container finished" podID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerID="5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738" exitCode=0 Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.254824 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdvpp" event={"ID":"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee","Type":"ContainerDied","Data":"5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.254874 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdvpp" event={"ID":"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee","Type":"ContainerStarted","Data":"3b2857ba49a15d2f148e5669d25e0da8f7bab92660e4d7b6928825a0a948d6d0"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.267169 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" event={"ID":"2b9bc3f6-3229-4432-ac8c-ef9395664f67","Type":"ContainerStarted","Data":"5ac4d5100ba3d363ae301ea8dc3da4137351194cd126118fb71eb7076519bc1c"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.330127 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vnn24"] Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.332084 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.337399 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.338524 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hrhlh" event={"ID":"1fa70255-2b2f-4df4-a080-e5fa22f847dd","Type":"ContainerStarted","Data":"5a123c1697439562fbd97a7d6e9d174772a8689c5579b11145115c6b1ef3b518"} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.368670 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.369610 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.374189 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.874163954 +0000 UTC m=+163.669011746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.393133 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vnn24"] Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.393535 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" podStartSLOduration=132.393519135 podStartE2EDuration="2m12.393519135s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:00.362737554 +0000 UTC m=+163.157585346" watchObservedRunningTime="2025-12-10 18:58:00.393519135 +0000 UTC m=+163.188366927" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.422505 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5db7f" podStartSLOduration=132.422467298 podStartE2EDuration="2m12.422467298s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:00.392377199 +0000 UTC m=+163.187224991" watchObservedRunningTime="2025-12-10 18:58:00.422467298 +0000 UTC m=+163.217315080" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.477383 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.477972 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-catalog-content\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.478076 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk8l4\" (UniqueName: \"kubernetes.io/projected/a5e7b198-78b1-41d4-86e9-583e68ae92f1-kube-api-access-pk8l4\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.478308 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-utilities\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.479102 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.979088544 +0000 UTC m=+163.773936326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.579809 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-catalog-content\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.579873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk8l4\" (UniqueName: \"kubernetes.io/projected/a5e7b198-78b1-41d4-86e9-583e68ae92f1-kube-api-access-pk8l4\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.579910 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.579928 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-utilities\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.580308 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-utilities\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.580502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-catalog-content\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.580939 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:58:01.080928757 +0000 UTC m=+163.875776549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jnckv" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.632065 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk8l4\" (UniqueName: \"kubernetes.io/projected/a5e7b198-78b1-41d4-86e9-583e68ae92f1-kube-api-access-pk8l4\") pod \"redhat-operators-vnn24\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.664473 4894 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.675902 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.681101 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:58:00 crc kubenswrapper[4894]: E1210 18:58:00.681408 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:58:01.181393967 +0000 UTC m=+163.976241759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.710805 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kt5vd"] Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.734622 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tdxmd"] Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.735833 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.739218 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.742085 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.742296 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tdxmd"] Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.753944 4894 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-10T18:58:00.664742881Z","Handler":null,"Name":""} Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.764723 4894 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.764757 4894 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.785514 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.813351 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.813389 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.868646 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jnckv\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.890622 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.891029 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-utilities\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.891111 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q255q\" (UniqueName: \"kubernetes.io/projected/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-kube-api-access-q255q\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.891217 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-catalog-content\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.964383 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.994549 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-catalog-content\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.994637 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-utilities\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.994673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q255q\" (UniqueName: \"kubernetes.io/projected/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-kube-api-access-q255q\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.995333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-utilities\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:00 crc kubenswrapper[4894]: I1210 18:58:00.995557 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-catalog-content\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.030927 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q255q\" (UniqueName: \"kubernetes.io/projected/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-kube-api-access-q255q\") pod \"redhat-operators-tdxmd\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.076833 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vnn24"] Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.081153 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.100140 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.146465 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.216993 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:01 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:01 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:01 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.217043 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.347323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" event={"ID":"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b","Type":"ContainerStarted","Data":"440cdd5481cf563650a922fb26bb7c713f604b015cc26062b51903ccc40d24cd"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.347362 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" event={"ID":"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b","Type":"ContainerStarted","Data":"28fdabb6f0e8e6aebbb5fdceba5c4d7e990f1e9f46f6f7d791e50a9287529a20"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.348136 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerStarted","Data":"ac03c2aaebbe519b9d6bc0a8b73b7c8e84c014eb63e4d2c773c72d3a58881fc4"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.349397 4894 generic.go:334] "Generic (PLEG): container finished" podID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerID="81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a" exitCode=0 Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.349441 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mb95" event={"ID":"91c2b024-f9b4-4d85-bc5d-ad463ae51871","Type":"ContainerDied","Data":"81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.349456 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mb95" event={"ID":"91c2b024-f9b4-4d85-bc5d-ad463ae51871","Type":"ContainerStarted","Data":"e2ec6c230fa29dd389472d457c92c6a3e7e02da42c3c1225a7fbcb83445b468d"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.354049 4894 generic.go:334] "Generic (PLEG): container finished" podID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerID="eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec" exitCode=0 Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.355468 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerDied","Data":"eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.355498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerStarted","Data":"35ea39d0807d6860b7da8143caba2f031501699749621ddd091d794fbf637912"} Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.468434 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.551498 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.772690 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tdxmd"] Dec 10 18:58:01 crc kubenswrapper[4894]: W1210 18:58:01.780326 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2badda9c_032f_49d5_9ed1_8d8768e7c1ef.slice/crio-1d86a005c22994665f26d190eed17818d1dca06c7bc1bf2fa5407d92dba787dd WatchSource:0}: Error finding container 1d86a005c22994665f26d190eed17818d1dca06c7bc1bf2fa5407d92dba787dd: Status 404 returned error can't find the container with id 1d86a005c22994665f26d190eed17818d1dca06c7bc1bf2fa5407d92dba787dd Dec 10 18:58:01 crc kubenswrapper[4894]: I1210 18:58:01.837103 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jnckv"] Dec 10 18:58:01 crc kubenswrapper[4894]: W1210 18:58:01.857266 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd03795b4_dfb1_47b7_9cfe_17f03c460f37.slice/crio-da3c81cce41537d54e696bb310ca0dd32d0b7383a07c6b95afe8dbcf04218d5e WatchSource:0}: Error finding container da3c81cce41537d54e696bb310ca0dd32d0b7383a07c6b95afe8dbcf04218d5e: Status 404 returned error can't find the container with id da3c81cce41537d54e696bb310ca0dd32d0b7383a07c6b95afe8dbcf04218d5e Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.211180 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:02 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:02 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:02 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.211244 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.361043 4894 generic.go:334] "Generic (PLEG): container finished" podID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerID="4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f" exitCode=0 Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.361110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerDied","Data":"4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.361141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerStarted","Data":"1d86a005c22994665f26d190eed17818d1dca06c7bc1bf2fa5407d92dba787dd"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.367279 4894 generic.go:334] "Generic (PLEG): container finished" podID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerID="364da9cc0eb2f9e763d018fccbf352e53f2eafecf5ec65d99a9b101acdc919d5" exitCode=0 Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.367348 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerDied","Data":"364da9cc0eb2f9e763d018fccbf352e53f2eafecf5ec65d99a9b101acdc919d5"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.370144 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" event={"ID":"d03795b4-dfb1-47b7-9cfe-17f03c460f37","Type":"ContainerStarted","Data":"0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.370186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" event={"ID":"d03795b4-dfb1-47b7-9cfe-17f03c460f37","Type":"ContainerStarted","Data":"da3c81cce41537d54e696bb310ca0dd32d0b7383a07c6b95afe8dbcf04218d5e"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.370722 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.373616 4894 generic.go:334] "Generic (PLEG): container finished" podID="e39c4e75-61aa-4690-bff2-eb4be173b66e" containerID="1b8571e70614c83cfc2f4179a1b158b448476733f2e469dce6ecac61f18f41fd" exitCode=0 Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.373702 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" event={"ID":"e39c4e75-61aa-4690-bff2-eb4be173b66e","Type":"ContainerDied","Data":"1b8571e70614c83cfc2f4179a1b158b448476733f2e469dce6ecac61f18f41fd"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.376115 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" event={"ID":"8efb3752-ad6a-4a44-8ac6-f5bc9bdd842b","Type":"ContainerStarted","Data":"446f7da635dd09c96ea597cf0fdb9b711f57ee9e507d25b99d35a77c29ebf956"} Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.388652 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t2p24" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.419426 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jkvhm" podStartSLOduration=11.419404946 podStartE2EDuration="11.419404946s" podCreationTimestamp="2025-12-10 18:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:02.417526697 +0000 UTC m=+165.212374489" watchObservedRunningTime="2025-12-10 18:58:02.419404946 +0000 UTC m=+165.214252748" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.464759 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" podStartSLOduration=134.464738496 podStartE2EDuration="2m14.464738496s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:02.450370482 +0000 UTC m=+165.245218274" watchObservedRunningTime="2025-12-10 18:58:02.464738496 +0000 UTC m=+165.259586288" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.597322 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bqdvr" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.792038 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:58:02 crc kubenswrapper[4894]: I1210 18:58:02.795878 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6pnv8" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.210919 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:03 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:03 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:03 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.211391 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.615340 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.762938 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6gl9\" (UniqueName: \"kubernetes.io/projected/e39c4e75-61aa-4690-bff2-eb4be173b66e-kube-api-access-c6gl9\") pod \"e39c4e75-61aa-4690-bff2-eb4be173b66e\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.763575 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e39c4e75-61aa-4690-bff2-eb4be173b66e-secret-volume\") pod \"e39c4e75-61aa-4690-bff2-eb4be173b66e\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.763607 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume\") pod \"e39c4e75-61aa-4690-bff2-eb4be173b66e\" (UID: \"e39c4e75-61aa-4690-bff2-eb4be173b66e\") " Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.764287 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume" (OuterVolumeSpecName: "config-volume") pod "e39c4e75-61aa-4690-bff2-eb4be173b66e" (UID: "e39c4e75-61aa-4690-bff2-eb4be173b66e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.779343 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e39c4e75-61aa-4690-bff2-eb4be173b66e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e39c4e75-61aa-4690-bff2-eb4be173b66e" (UID: "e39c4e75-61aa-4690-bff2-eb4be173b66e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.786810 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e39c4e75-61aa-4690-bff2-eb4be173b66e-kube-api-access-c6gl9" (OuterVolumeSpecName: "kube-api-access-c6gl9") pod "e39c4e75-61aa-4690-bff2-eb4be173b66e" (UID: "e39c4e75-61aa-4690-bff2-eb4be173b66e"). InnerVolumeSpecName "kube-api-access-c6gl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.865093 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e39c4e75-61aa-4690-bff2-eb4be173b66e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.865126 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e39c4e75-61aa-4690-bff2-eb4be173b66e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:03 crc kubenswrapper[4894]: I1210 18:58:03.865140 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6gl9\" (UniqueName: \"kubernetes.io/projected/e39c4e75-61aa-4690-bff2-eb4be173b66e-kube-api-access-c6gl9\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.208147 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.210118 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:04 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:04 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:04 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.210170 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.392157 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.392185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n" event={"ID":"e39c4e75-61aa-4690-bff2-eb4be173b66e","Type":"ContainerDied","Data":"e7cab4bcc90af69816eb6e943821655e19bb2bc58b1427e61b39bc0e69ff7bab"} Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.392894 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7cab4bcc90af69816eb6e943821655e19bb2bc58b1427e61b39bc0e69ff7bab" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.435627 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 18:58:04 crc kubenswrapper[4894]: E1210 18:58:04.435902 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39c4e75-61aa-4690-bff2-eb4be173b66e" containerName="collect-profiles" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.435914 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39c4e75-61aa-4690-bff2-eb4be173b66e" containerName="collect-profiles" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.436017 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39c4e75-61aa-4690-bff2-eb4be173b66e" containerName="collect-profiles" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.436383 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.438273 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.439122 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.439173 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.578834 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.579200 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.596305 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.680585 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.680657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.681502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.699378 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:04 crc kubenswrapper[4894]: I1210 18:58:04.767723 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.128716 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.212586 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:05 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:05 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:05 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.212639 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.408516 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48b5df9d-e145-4fcb-8ee5-b860de1ffa99","Type":"ContainerStarted","Data":"c635f95431a38be8c1ac9376e7560c7c95933aa7590d99f216fb09e6f5be80bc"} Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.645148 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.645622 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.647923 4894 patch_prober.go:28] interesting pod/console-f9d7485db-vsms7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 10 18:58:05 crc kubenswrapper[4894]: I1210 18:58:05.648048 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vsms7" podUID="139baf99-78f8-41f5-8c3c-2d37c42edd47" containerName="console" probeResult="failure" output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 10 18:58:06 crc kubenswrapper[4894]: I1210 18:58:06.210566 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:06 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:06 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:06 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:06 crc kubenswrapper[4894]: I1210 18:58:06.210812 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:06 crc kubenswrapper[4894]: I1210 18:58:06.415123 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48b5df9d-e145-4fcb-8ee5-b860de1ffa99","Type":"ContainerStarted","Data":"b309ec94816a5b3cae6d28bff9c227451d6229e869c053df3f0914429c03dbe3"} Dec 10 18:58:06 crc kubenswrapper[4894]: I1210 18:58:06.428406 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.428387417 podStartE2EDuration="2.428387417s" podCreationTimestamp="2025-12-10 18:58:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:06.424787114 +0000 UTC m=+169.219634926" watchObservedRunningTime="2025-12-10 18:58:06.428387417 +0000 UTC m=+169.223235209" Dec 10 18:58:06 crc kubenswrapper[4894]: I1210 18:58:06.758885 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hm4c5" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.214589 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:07 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:07 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:07 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.214650 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.354147 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.355170 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.358202 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.361620 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.364767 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.431095 4894 generic.go:334] "Generic (PLEG): container finished" podID="48b5df9d-e145-4fcb-8ee5-b860de1ffa99" containerID="b309ec94816a5b3cae6d28bff9c227451d6229e869c053df3f0914429c03dbe3" exitCode=0 Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.431135 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48b5df9d-e145-4fcb-8ee5-b860de1ffa99","Type":"ContainerDied","Data":"b309ec94816a5b3cae6d28bff9c227451d6229e869c053df3f0914429c03dbe3"} Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.431578 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26a358e6-4be8-4867-8307-b685dae13d6f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.431655 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26a358e6-4be8-4867-8307-b685dae13d6f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.532873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26a358e6-4be8-4867-8307-b685dae13d6f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.532939 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26a358e6-4be8-4867-8307-b685dae13d6f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.533011 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26a358e6-4be8-4867-8307-b685dae13d6f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.569633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26a358e6-4be8-4867-8307-b685dae13d6f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.703272 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:07 crc kubenswrapper[4894]: I1210 18:58:07.890635 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.211096 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:08 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:08 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:08 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.211367 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.436913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"26a358e6-4be8-4867-8307-b685dae13d6f","Type":"ContainerStarted","Data":"23e38ca46a331dfec9ce4464d6beff89de706130966b61c378ad749d6075163d"} Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.723090 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.850407 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kube-api-access\") pod \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.850475 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kubelet-dir\") pod \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\" (UID: \"48b5df9d-e145-4fcb-8ee5-b860de1ffa99\") " Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.850599 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "48b5df9d-e145-4fcb-8ee5-b860de1ffa99" (UID: "48b5df9d-e145-4fcb-8ee5-b860de1ffa99"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.850777 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.859306 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "48b5df9d-e145-4fcb-8ee5-b860de1ffa99" (UID: "48b5df9d-e145-4fcb-8ee5-b860de1ffa99"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:08 crc kubenswrapper[4894]: I1210 18:58:08.952173 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48b5df9d-e145-4fcb-8ee5-b860de1ffa99-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:09 crc kubenswrapper[4894]: I1210 18:58:09.211639 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:09 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:09 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:09 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:09 crc kubenswrapper[4894]: I1210 18:58:09.211706 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:09 crc kubenswrapper[4894]: I1210 18:58:09.442777 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"26a358e6-4be8-4867-8307-b685dae13d6f","Type":"ContainerStarted","Data":"855c1217ff1fb006b2c185f32104a75d47532477e4bc5ae9f4607e2a36ec2cb8"} Dec 10 18:58:09 crc kubenswrapper[4894]: I1210 18:58:09.444315 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"48b5df9d-e145-4fcb-8ee5-b860de1ffa99","Type":"ContainerDied","Data":"c635f95431a38be8c1ac9376e7560c7c95933aa7590d99f216fb09e6f5be80bc"} Dec 10 18:58:09 crc kubenswrapper[4894]: I1210 18:58:09.444333 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c635f95431a38be8c1ac9376e7560c7c95933aa7590d99f216fb09e6f5be80bc" Dec 10 18:58:09 crc kubenswrapper[4894]: I1210 18:58:09.444408 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:58:10 crc kubenswrapper[4894]: I1210 18:58:10.210943 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:10 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:10 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:10 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:10 crc kubenswrapper[4894]: I1210 18:58:10.211055 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.089013 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.103276 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72521f10-d803-4b10-97dc-cbad9bda6be7-metrics-certs\") pod \"network-metrics-daemon-7dks9\" (UID: \"72521f10-d803-4b10-97dc-cbad9bda6be7\") " pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.184898 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dks9" Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.211552 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:11 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 10 18:58:11 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:11 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.211874 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.455413 4894 generic.go:334] "Generic (PLEG): container finished" podID="26a358e6-4be8-4867-8307-b685dae13d6f" containerID="855c1217ff1fb006b2c185f32104a75d47532477e4bc5ae9f4607e2a36ec2cb8" exitCode=0 Dec 10 18:58:11 crc kubenswrapper[4894]: I1210 18:58:11.455460 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"26a358e6-4be8-4867-8307-b685dae13d6f","Type":"ContainerDied","Data":"855c1217ff1fb006b2c185f32104a75d47532477e4bc5ae9f4607e2a36ec2cb8"} Dec 10 18:58:12 crc kubenswrapper[4894]: I1210 18:58:12.214566 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:58:12 crc kubenswrapper[4894]: [+]has-synced ok Dec 10 18:58:12 crc kubenswrapper[4894]: [+]process-running ok Dec 10 18:58:12 crc kubenswrapper[4894]: healthz check failed Dec 10 18:58:12 crc kubenswrapper[4894]: I1210 18:58:12.214947 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:58:13 crc kubenswrapper[4894]: I1210 18:58:13.216590 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:58:13 crc kubenswrapper[4894]: I1210 18:58:13.220383 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mk6q8" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.109491 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.171727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26a358e6-4be8-4867-8307-b685dae13d6f-kubelet-dir\") pod \"26a358e6-4be8-4867-8307-b685dae13d6f\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.171838 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a358e6-4be8-4867-8307-b685dae13d6f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "26a358e6-4be8-4867-8307-b685dae13d6f" (UID: "26a358e6-4be8-4867-8307-b685dae13d6f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.171985 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26a358e6-4be8-4867-8307-b685dae13d6f-kube-api-access\") pod \"26a358e6-4be8-4867-8307-b685dae13d6f\" (UID: \"26a358e6-4be8-4867-8307-b685dae13d6f\") " Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.172299 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26a358e6-4be8-4867-8307-b685dae13d6f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.176996 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a358e6-4be8-4867-8307-b685dae13d6f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "26a358e6-4be8-4867-8307-b685dae13d6f" (UID: "26a358e6-4be8-4867-8307-b685dae13d6f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.273680 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26a358e6-4be8-4867-8307-b685dae13d6f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.481242 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"26a358e6-4be8-4867-8307-b685dae13d6f","Type":"ContainerDied","Data":"23e38ca46a331dfec9ce4464d6beff89de706130966b61c378ad749d6075163d"} Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.481288 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23e38ca46a331dfec9ce4464d6beff89de706130966b61c378ad749d6075163d" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.481318 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.648377 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:58:15 crc kubenswrapper[4894]: I1210 18:58:15.651454 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 18:58:19 crc kubenswrapper[4894]: I1210 18:58:19.576770 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 18:58:19 crc kubenswrapper[4894]: I1210 18:58:19.577486 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 18:58:21 crc kubenswrapper[4894]: I1210 18:58:21.093157 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 18:58:31 crc kubenswrapper[4894]: I1210 18:58:31.102548 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:58:31 crc kubenswrapper[4894]: E1210 18:58:31.945987 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 18:58:31 crc kubenswrapper[4894]: E1210 18:58:31.946271 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mzbw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zdvpp_openshift-marketplace(b845e384-badb-4fbe-92c5-9ba9b9b8a8ee): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:31 crc kubenswrapper[4894]: E1210 18:58:31.947727 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zdvpp" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" Dec 10 18:58:34 crc kubenswrapper[4894]: E1210 18:58:34.818144 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zdvpp" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" Dec 10 18:58:35 crc kubenswrapper[4894]: E1210 18:58:35.078189 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 18:58:35 crc kubenswrapper[4894]: E1210 18:58:35.078358 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqxz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-m7tz9_openshift-marketplace(3a0cf23c-9387-4ed5-8f07-95b54c449786): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:35 crc kubenswrapper[4894]: E1210 18:58:35.079606 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-m7tz9" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" Dec 10 18:58:35 crc kubenswrapper[4894]: I1210 18:58:35.529014 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kchjv" Dec 10 18:58:35 crc kubenswrapper[4894]: E1210 18:58:35.891063 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 18:58:35 crc kubenswrapper[4894]: E1210 18:58:35.891222 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljt58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dkr5p_openshift-marketplace(dada86bd-f99d-4566-98eb-33194725df47): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:35 crc kubenswrapper[4894]: E1210 18:58:35.893075 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dkr5p" podUID="dada86bd-f99d-4566-98eb-33194725df47" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.736749 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dkr5p" podUID="dada86bd-f99d-4566-98eb-33194725df47" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.737222 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-m7tz9" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.805668 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.806011 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jvwcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kt5vd_openshift-marketplace(f8cb70e6-c927-4dea-8184-ae0a5343ddec): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.807166 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kt5vd" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.815857 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.816019 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5867z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9mb95_openshift-marketplace(91c2b024-f9b4-4d85-bc5d-ad463ae51871): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.817382 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-9mb95" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.822519 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.822777 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cb52m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vr2zg_openshift-marketplace(2a2f64bf-a1f8-4ad3-9dd3-970afc101879): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:37 crc kubenswrapper[4894]: E1210 18:58:37.824401 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vr2zg" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.754549 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 18:58:38 crc kubenswrapper[4894]: E1210 18:58:38.754905 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a358e6-4be8-4867-8307-b685dae13d6f" containerName="pruner" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.754922 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a358e6-4be8-4867-8307-b685dae13d6f" containerName="pruner" Dec 10 18:58:38 crc kubenswrapper[4894]: E1210 18:58:38.754935 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b5df9d-e145-4fcb-8ee5-b860de1ffa99" containerName="pruner" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.754944 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b5df9d-e145-4fcb-8ee5-b860de1ffa99" containerName="pruner" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.755072 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a358e6-4be8-4867-8307-b685dae13d6f" containerName="pruner" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.755094 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b5df9d-e145-4fcb-8ee5-b860de1ffa99" containerName="pruner" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.755580 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.758145 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.759435 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.760872 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.843357 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.843612 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.944627 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.944672 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.944789 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:38 crc kubenswrapper[4894]: I1210 18:58:38.963464 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:39 crc kubenswrapper[4894]: I1210 18:58:39.075143 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:41 crc kubenswrapper[4894]: E1210 18:58:41.054774 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kt5vd" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" Dec 10 18:58:41 crc kubenswrapper[4894]: E1210 18:58:41.054830 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9mb95" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" Dec 10 18:58:41 crc kubenswrapper[4894]: E1210 18:58:41.054944 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vr2zg" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.283210 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7dks9"] Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.342984 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 18:58:41 crc kubenswrapper[4894]: W1210 18:58:41.351149 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3b53e16f_8f52_48f3_bb9c_c018d2da6c48.slice/crio-a9eec5a1a2fe5503a91b46ea665f7159fec75eed1c9615e57cb8652d85586f97 WatchSource:0}: Error finding container a9eec5a1a2fe5503a91b46ea665f7159fec75eed1c9615e57cb8652d85586f97: Status 404 returned error can't find the container with id a9eec5a1a2fe5503a91b46ea665f7159fec75eed1c9615e57cb8652d85586f97 Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.647494 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerStarted","Data":"c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9"} Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.651719 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7dks9" event={"ID":"72521f10-d803-4b10-97dc-cbad9bda6be7","Type":"ContainerStarted","Data":"e5c684d9a6e0a723ed68da52cc47e3949d52a52284e79ed1e20b92159d1d7326"} Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.651763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7dks9" event={"ID":"72521f10-d803-4b10-97dc-cbad9bda6be7","Type":"ContainerStarted","Data":"7932f1e82bcafd07b37d382f613d95786f988b9d535f1e7b2265174948775e49"} Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.655366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerStarted","Data":"2631b3936979dfc11e34d3fa0a5b685ad3958937e258efe91ac8468c62201d91"} Dec 10 18:58:41 crc kubenswrapper[4894]: I1210 18:58:41.657461 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3b53e16f-8f52-48f3-bb9c-c018d2da6c48","Type":"ContainerStarted","Data":"a9eec5a1a2fe5503a91b46ea665f7159fec75eed1c9615e57cb8652d85586f97"} Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.664390 4894 generic.go:334] "Generic (PLEG): container finished" podID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerID="c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9" exitCode=0 Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.664803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerDied","Data":"c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9"} Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.669671 4894 generic.go:334] "Generic (PLEG): container finished" podID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerID="2631b3936979dfc11e34d3fa0a5b685ad3958937e258efe91ac8468c62201d91" exitCode=0 Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.669709 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerDied","Data":"2631b3936979dfc11e34d3fa0a5b685ad3958937e258efe91ac8468c62201d91"} Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.671718 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7dks9" event={"ID":"72521f10-d803-4b10-97dc-cbad9bda6be7","Type":"ContainerStarted","Data":"badbb0e51820a632ef0505da1325017b775c27f26de422a8d42dcc7c08e18ac8"} Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.675666 4894 generic.go:334] "Generic (PLEG): container finished" podID="3b53e16f-8f52-48f3-bb9c-c018d2da6c48" containerID="5583d82d16dd41a866c29b8752caa12de6b161cdbe7e483a927e49269b748fa2" exitCode=0 Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.675688 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3b53e16f-8f52-48f3-bb9c-c018d2da6c48","Type":"ContainerDied","Data":"5583d82d16dd41a866c29b8752caa12de6b161cdbe7e483a927e49269b748fa2"} Dec 10 18:58:42 crc kubenswrapper[4894]: I1210 18:58:42.708189 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7dks9" podStartSLOduration=174.708165564 podStartE2EDuration="2m54.708165564s" podCreationTimestamp="2025-12-10 18:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:42.707054932 +0000 UTC m=+205.501902734" watchObservedRunningTime="2025-12-10 18:58:42.708165564 +0000 UTC m=+205.503013356" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.555276 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.556394 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.565547 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.628740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kube-api-access\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.628814 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.628927 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-var-lock\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.682346 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerStarted","Data":"fa3bd2137101ef56eed2e6b51530d50e3b3b39ba5335c6e7eb7f923dc0bb991c"} Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.687134 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerStarted","Data":"ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992"} Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.703016 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vnn24" podStartSLOduration=2.64257153 podStartE2EDuration="43.702993554s" podCreationTimestamp="2025-12-10 18:58:00 +0000 UTC" firstStartedPulling="2025-12-10 18:58:02.369196092 +0000 UTC m=+165.164043884" lastFinishedPulling="2025-12-10 18:58:43.429618116 +0000 UTC m=+206.224465908" observedRunningTime="2025-12-10 18:58:43.700359139 +0000 UTC m=+206.495206941" watchObservedRunningTime="2025-12-10 18:58:43.702993554 +0000 UTC m=+206.497841356" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.722109 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tdxmd" podStartSLOduration=2.66644564 podStartE2EDuration="43.722090262s" podCreationTimestamp="2025-12-10 18:58:00 +0000 UTC" firstStartedPulling="2025-12-10 18:58:02.362448739 +0000 UTC m=+165.157296531" lastFinishedPulling="2025-12-10 18:58:43.418093361 +0000 UTC m=+206.212941153" observedRunningTime="2025-12-10 18:58:43.721217737 +0000 UTC m=+206.516065559" watchObservedRunningTime="2025-12-10 18:58:43.722090262 +0000 UTC m=+206.516938054" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.729686 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kube-api-access\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.729733 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.729772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-var-lock\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.729831 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-var-lock\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.729883 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.753741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kube-api-access\") pod \"installer-9-crc\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.876290 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:43 crc kubenswrapper[4894]: I1210 18:58:43.929569 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.033421 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kubelet-dir\") pod \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.033589 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kube-api-access\") pod \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\" (UID: \"3b53e16f-8f52-48f3-bb9c-c018d2da6c48\") " Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.033581 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3b53e16f-8f52-48f3-bb9c-c018d2da6c48" (UID: "3b53e16f-8f52-48f3-bb9c-c018d2da6c48"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.033796 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.038830 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3b53e16f-8f52-48f3-bb9c-c018d2da6c48" (UID: "3b53e16f-8f52-48f3-bb9c-c018d2da6c48"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.134728 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b53e16f-8f52-48f3-bb9c-c018d2da6c48-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.268182 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 18:58:44 crc kubenswrapper[4894]: W1210 18:58:44.280974 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode5bc823d_aae9_4d4e_b6cd_095590924c6c.slice/crio-e0c64a1098e0246964ab6c40ae84f25d623d32af8421bba8ec4c03e489e08da2 WatchSource:0}: Error finding container e0c64a1098e0246964ab6c40ae84f25d623d32af8421bba8ec4c03e489e08da2: Status 404 returned error can't find the container with id e0c64a1098e0246964ab6c40ae84f25d623d32af8421bba8ec4c03e489e08da2 Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.699446 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e5bc823d-aae9-4d4e-b6cd-095590924c6c","Type":"ContainerStarted","Data":"a8982172dfc36ee4653ff5b0182fd7d5699f08fbc017842c7684cdd1f1d83ffe"} Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.699496 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e5bc823d-aae9-4d4e-b6cd-095590924c6c","Type":"ContainerStarted","Data":"e0c64a1098e0246964ab6c40ae84f25d623d32af8421bba8ec4c03e489e08da2"} Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.702926 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.707024 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3b53e16f-8f52-48f3-bb9c-c018d2da6c48","Type":"ContainerDied","Data":"a9eec5a1a2fe5503a91b46ea665f7159fec75eed1c9615e57cb8652d85586f97"} Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.707054 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9eec5a1a2fe5503a91b46ea665f7159fec75eed1c9615e57cb8652d85586f97" Dec 10 18:58:44 crc kubenswrapper[4894]: I1210 18:58:44.720798 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.720779851 podStartE2EDuration="1.720779851s" podCreationTimestamp="2025-12-10 18:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:44.718008723 +0000 UTC m=+207.512856535" watchObservedRunningTime="2025-12-10 18:58:44.720779851 +0000 UTC m=+207.515627643" Dec 10 18:58:49 crc kubenswrapper[4894]: I1210 18:58:49.577382 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 18:58:49 crc kubenswrapper[4894]: I1210 18:58:49.577919 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 18:58:49 crc kubenswrapper[4894]: I1210 18:58:49.578112 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 18:58:49 crc kubenswrapper[4894]: I1210 18:58:49.578836 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 18:58:49 crc kubenswrapper[4894]: I1210 18:58:49.579088 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955" gracePeriod=600 Dec 10 18:58:50 crc kubenswrapper[4894]: I1210 18:58:50.677437 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:50 crc kubenswrapper[4894]: I1210 18:58:50.678677 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.078116 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.101200 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.101279 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.139658 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.741354 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955" exitCode=0 Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.741444 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955"} Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.794707 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 18:58:51 crc kubenswrapper[4894]: I1210 18:58:51.804177 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:52 crc kubenswrapper[4894]: I1210 18:58:52.755319 4894 generic.go:334] "Generic (PLEG): container finished" podID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerID="62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a" exitCode=0 Dec 10 18:58:52 crc kubenswrapper[4894]: I1210 18:58:52.755408 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdvpp" event={"ID":"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee","Type":"ContainerDied","Data":"62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a"} Dec 10 18:58:52 crc kubenswrapper[4894]: I1210 18:58:52.761820 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"a67b89ac2990b2f668b7964537a50bb5050a9050493de8896a0931aefdc247e9"} Dec 10 18:58:52 crc kubenswrapper[4894]: I1210 18:58:52.766253 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerStarted","Data":"78455d4420edbfeb783b49166aa9dea7eb2387c74fbda2dce34818bdd6b7ad1d"} Dec 10 18:58:52 crc kubenswrapper[4894]: I1210 18:58:52.770124 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tdxmd"] Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.772625 4894 generic.go:334] "Generic (PLEG): container finished" podID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerID="0fbc5bd9d991bc1189bd4818a253c28450d111f9d232bdfe32ba06136a6c218a" exitCode=0 Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.772731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7tz9" event={"ID":"3a0cf23c-9387-4ed5-8f07-95b54c449786","Type":"ContainerDied","Data":"0fbc5bd9d991bc1189bd4818a253c28450d111f9d232bdfe32ba06136a6c218a"} Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.778881 4894 generic.go:334] "Generic (PLEG): container finished" podID="dada86bd-f99d-4566-98eb-33194725df47" containerID="78455d4420edbfeb783b49166aa9dea7eb2387c74fbda2dce34818bdd6b7ad1d" exitCode=0 Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.778962 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerDied","Data":"78455d4420edbfeb783b49166aa9dea7eb2387c74fbda2dce34818bdd6b7ad1d"} Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.779054 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tdxmd" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="registry-server" containerID="cri-o://ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992" gracePeriod=2 Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.822330 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hqtqc"] Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.822563 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" podUID="7fe4fbd5-317b-49fd-9c30-82860839a273" containerName="controller-manager" containerID="cri-o://89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b" gracePeriod=30 Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.919116 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r"] Dec 10 18:58:53 crc kubenswrapper[4894]: I1210 18:58:53.919668 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" podUID="89fbf7f8-3630-4587-95da-e8bf12127a64" containerName="route-controller-manager" containerID="cri-o://b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d" gracePeriod=30 Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.161570 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.245177 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.273689 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe4fbd5-317b-49fd-9c30-82860839a273-serving-cert\") pod \"7fe4fbd5-317b-49fd-9c30-82860839a273\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.273948 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-client-ca\") pod \"7fe4fbd5-317b-49fd-9c30-82860839a273\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.274063 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-config\") pod \"7fe4fbd5-317b-49fd-9c30-82860839a273\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.274178 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7h5q\" (UniqueName: \"kubernetes.io/projected/7fe4fbd5-317b-49fd-9c30-82860839a273-kube-api-access-d7h5q\") pod \"7fe4fbd5-317b-49fd-9c30-82860839a273\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.274285 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-proxy-ca-bundles\") pod \"7fe4fbd5-317b-49fd-9c30-82860839a273\" (UID: \"7fe4fbd5-317b-49fd-9c30-82860839a273\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.275553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7fe4fbd5-317b-49fd-9c30-82860839a273" (UID: "7fe4fbd5-317b-49fd-9c30-82860839a273"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.276334 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-client-ca" (OuterVolumeSpecName: "client-ca") pod "7fe4fbd5-317b-49fd-9c30-82860839a273" (UID: "7fe4fbd5-317b-49fd-9c30-82860839a273"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.277275 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-config" (OuterVolumeSpecName: "config") pod "7fe4fbd5-317b-49fd-9c30-82860839a273" (UID: "7fe4fbd5-317b-49fd-9c30-82860839a273"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.281724 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe4fbd5-317b-49fd-9c30-82860839a273-kube-api-access-d7h5q" (OuterVolumeSpecName: "kube-api-access-d7h5q") pod "7fe4fbd5-317b-49fd-9c30-82860839a273" (UID: "7fe4fbd5-317b-49fd-9c30-82860839a273"). InnerVolumeSpecName "kube-api-access-d7h5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.281962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fe4fbd5-317b-49fd-9c30-82860839a273-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7fe4fbd5-317b-49fd-9c30-82860839a273" (UID: "7fe4fbd5-317b-49fd-9c30-82860839a273"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.375624 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") pod \"89fbf7f8-3630-4587-95da-e8bf12127a64\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.375673 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") pod \"89fbf7f8-3630-4587-95da-e8bf12127a64\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.375701 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfk4x\" (UniqueName: \"kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x\") pod \"89fbf7f8-3630-4587-95da-e8bf12127a64\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.375771 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") pod \"89fbf7f8-3630-4587-95da-e8bf12127a64\" (UID: \"89fbf7f8-3630-4587-95da-e8bf12127a64\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376034 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376051 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe4fbd5-317b-49fd-9c30-82860839a273-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376059 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376069 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4fbd5-317b-49fd-9c30-82860839a273-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376078 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7h5q\" (UniqueName: \"kubernetes.io/projected/7fe4fbd5-317b-49fd-9c30-82860839a273-kube-api-access-d7h5q\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376757 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca" (OuterVolumeSpecName: "client-ca") pod "89fbf7f8-3630-4587-95da-e8bf12127a64" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.376804 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config" (OuterVolumeSpecName: "config") pod "89fbf7f8-3630-4587-95da-e8bf12127a64" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.380280 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "89fbf7f8-3630-4587-95da-e8bf12127a64" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.380626 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x" (OuterVolumeSpecName: "kube-api-access-sfk4x") pod "89fbf7f8-3630-4587-95da-e8bf12127a64" (UID: "89fbf7f8-3630-4587-95da-e8bf12127a64"). InnerVolumeSpecName "kube-api-access-sfk4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.476863 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89fbf7f8-3630-4587-95da-e8bf12127a64-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.476920 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.476933 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfk4x\" (UniqueName: \"kubernetes.io/projected/89fbf7f8-3630-4587-95da-e8bf12127a64-kube-api-access-sfk4x\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.476944 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89fbf7f8-3630-4587-95da-e8bf12127a64-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.598002 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.678893 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-catalog-content\") pod \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.678960 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-utilities\") pod \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.679053 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q255q\" (UniqueName: \"kubernetes.io/projected/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-kube-api-access-q255q\") pod \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\" (UID: \"2badda9c-032f-49d5-9ed1-8d8768e7c1ef\") " Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.679875 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-utilities" (OuterVolumeSpecName: "utilities") pod "2badda9c-032f-49d5-9ed1-8d8768e7c1ef" (UID: "2badda9c-032f-49d5-9ed1-8d8768e7c1ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.681515 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-kube-api-access-q255q" (OuterVolumeSpecName: "kube-api-access-q255q") pod "2badda9c-032f-49d5-9ed1-8d8768e7c1ef" (UID: "2badda9c-032f-49d5-9ed1-8d8768e7c1ef"). InnerVolumeSpecName "kube-api-access-q255q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.779952 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q255q\" (UniqueName: \"kubernetes.io/projected/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-kube-api-access-q255q\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.780195 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.785802 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdvpp" event={"ID":"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee","Type":"ContainerStarted","Data":"13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.787519 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerStarted","Data":"9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.788955 4894 generic.go:334] "Generic (PLEG): container finished" podID="89fbf7f8-3630-4587-95da-e8bf12127a64" containerID="b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d" exitCode=0 Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.789013 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" event={"ID":"89fbf7f8-3630-4587-95da-e8bf12127a64","Type":"ContainerDied","Data":"b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.789039 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" event={"ID":"89fbf7f8-3630-4587-95da-e8bf12127a64","Type":"ContainerDied","Data":"66a16666c09f9120951338b28ff86c87742730c7f16b3c4d67e31dae212767b7"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.789053 4894 scope.go:117] "RemoveContainer" containerID="b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.789146 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.793862 4894 generic.go:334] "Generic (PLEG): container finished" podID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerID="ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992" exitCode=0 Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.793938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerDied","Data":"ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.793992 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdxmd" event={"ID":"2badda9c-032f-49d5-9ed1-8d8768e7c1ef","Type":"ContainerDied","Data":"1d86a005c22994665f26d190eed17818d1dca06c7bc1bf2fa5407d92dba787dd"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.794205 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdxmd" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.798396 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fe4fbd5-317b-49fd-9c30-82860839a273" containerID="89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b" exitCode=0 Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.798444 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" event={"ID":"7fe4fbd5-317b-49fd-9c30-82860839a273","Type":"ContainerDied","Data":"89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.798644 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" event={"ID":"7fe4fbd5-317b-49fd-9c30-82860839a273","Type":"ContainerDied","Data":"bbd4cec65ea5e1c3bbfd8cccbefae46217d614c6ea8ca21e50b35050aa69b25b"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.798473 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hqtqc" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.802547 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zdvpp" podStartSLOduration=4.364068337 podStartE2EDuration="57.802537637s" podCreationTimestamp="2025-12-10 18:57:57 +0000 UTC" firstStartedPulling="2025-12-10 18:58:00.266962513 +0000 UTC m=+163.061810305" lastFinishedPulling="2025-12-10 18:58:53.705431813 +0000 UTC m=+216.500279605" observedRunningTime="2025-12-10 18:58:54.802013772 +0000 UTC m=+217.596861584" watchObservedRunningTime="2025-12-10 18:58:54.802537637 +0000 UTC m=+217.597385429" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.803128 4894 generic.go:334] "Generic (PLEG): container finished" podID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerID="d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb" exitCode=0 Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.803206 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mb95" event={"ID":"91c2b024-f9b4-4d85-bc5d-ad463ae51871","Type":"ContainerDied","Data":"d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.804383 4894 scope.go:117] "RemoveContainer" containerID="b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d" Dec 10 18:58:54 crc kubenswrapper[4894]: E1210 18:58:54.805026 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d\": container with ID starting with b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d not found: ID does not exist" containerID="b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.805112 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d"} err="failed to get container status \"b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d\": rpc error: code = NotFound desc = could not find container \"b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d\": container with ID starting with b11bc78ab5b370c8e31d150138042b68bf0eef4f93fdd198ff0ab98b80edab9d not found: ID does not exist" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.805136 4894 scope.go:117] "RemoveContainer" containerID="ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.810061 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerStarted","Data":"db0aa1826da1f98708f7e6aee7db016fe6fcb7cc710455613cf3bc808c1ae3de"} Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.810709 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2badda9c-032f-49d5-9ed1-8d8768e7c1ef" (UID: "2badda9c-032f-49d5-9ed1-8d8768e7c1ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.839040 4894 scope.go:117] "RemoveContainer" containerID="c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.863700 4894 scope.go:117] "RemoveContainer" containerID="4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.878206 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dkr5p" podStartSLOduration=3.690731192 podStartE2EDuration="57.87818438s" podCreationTimestamp="2025-12-10 18:57:57 +0000 UTC" firstStartedPulling="2025-12-10 18:58:00.207268549 +0000 UTC m=+163.002116341" lastFinishedPulling="2025-12-10 18:58:54.394721737 +0000 UTC m=+217.189569529" observedRunningTime="2025-12-10 18:58:54.865390309 +0000 UTC m=+217.660238121" watchObservedRunningTime="2025-12-10 18:58:54.87818438 +0000 UTC m=+217.673032172" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.895600 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r"] Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.896571 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2badda9c-032f-49d5-9ed1-8d8768e7c1ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.897470 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m8r4r"] Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.900690 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hqtqc"] Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.903210 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hqtqc"] Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.928267 4894 scope.go:117] "RemoveContainer" containerID="ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992" Dec 10 18:58:54 crc kubenswrapper[4894]: E1210 18:58:54.928920 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992\": container with ID starting with ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992 not found: ID does not exist" containerID="ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.928954 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992"} err="failed to get container status \"ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992\": rpc error: code = NotFound desc = could not find container \"ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992\": container with ID starting with ff64a3e9907b75faffceaf6975ed2eaf48b721c308fa80fe25266c690e056992 not found: ID does not exist" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.928980 4894 scope.go:117] "RemoveContainer" containerID="c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9" Dec 10 18:58:54 crc kubenswrapper[4894]: E1210 18:58:54.929278 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9\": container with ID starting with c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9 not found: ID does not exist" containerID="c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.929292 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9"} err="failed to get container status \"c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9\": rpc error: code = NotFound desc = could not find container \"c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9\": container with ID starting with c8f13370a250d1dc057353314a482a9f971cbe1209358d2044d356a2e68d33e9 not found: ID does not exist" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.929303 4894 scope.go:117] "RemoveContainer" containerID="4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f" Dec 10 18:58:54 crc kubenswrapper[4894]: E1210 18:58:54.929454 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f\": container with ID starting with 4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f not found: ID does not exist" containerID="4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.929469 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f"} err="failed to get container status \"4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f\": rpc error: code = NotFound desc = could not find container \"4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f\": container with ID starting with 4a0316ce6d846a3c46fc3877f9601b3fac974adf576148ae5afd86958545189f not found: ID does not exist" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.929482 4894 scope.go:117] "RemoveContainer" containerID="89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.943759 4894 scope.go:117] "RemoveContainer" containerID="89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b" Dec 10 18:58:54 crc kubenswrapper[4894]: E1210 18:58:54.944105 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b\": container with ID starting with 89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b not found: ID does not exist" containerID="89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b" Dec 10 18:58:54 crc kubenswrapper[4894]: I1210 18:58:54.944143 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b"} err="failed to get container status \"89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b\": rpc error: code = NotFound desc = could not find container \"89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b\": container with ID starting with 89f8ca362a920cff697a03d290b5f9ff779c1139a371ac7168011929a2dffb7b not found: ID does not exist" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.120533 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tdxmd"] Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.127092 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tdxmd"] Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.547368 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" path="/var/lib/kubelet/pods/2badda9c-032f-49d5-9ed1-8d8768e7c1ef/volumes" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.548049 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe4fbd5-317b-49fd-9c30-82860839a273" path="/var/lib/kubelet/pods/7fe4fbd5-317b-49fd-9c30-82860839a273/volumes" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.548568 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89fbf7f8-3630-4587-95da-e8bf12127a64" path="/var/lib/kubelet/pods/89fbf7f8-3630-4587-95da-e8bf12127a64/volumes" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601451 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p"] Dec 10 18:58:55 crc kubenswrapper[4894]: E1210 18:58:55.601715 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="extract-content" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601732 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="extract-content" Dec 10 18:58:55 crc kubenswrapper[4894]: E1210 18:58:55.601745 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe4fbd5-317b-49fd-9c30-82860839a273" containerName="controller-manager" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601754 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe4fbd5-317b-49fd-9c30-82860839a273" containerName="controller-manager" Dec 10 18:58:55 crc kubenswrapper[4894]: E1210 18:58:55.601767 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fbf7f8-3630-4587-95da-e8bf12127a64" containerName="route-controller-manager" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601774 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fbf7f8-3630-4587-95da-e8bf12127a64" containerName="route-controller-manager" Dec 10 18:58:55 crc kubenswrapper[4894]: E1210 18:58:55.601789 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b53e16f-8f52-48f3-bb9c-c018d2da6c48" containerName="pruner" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601798 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b53e16f-8f52-48f3-bb9c-c018d2da6c48" containerName="pruner" Dec 10 18:58:55 crc kubenswrapper[4894]: E1210 18:58:55.601810 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="registry-server" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601818 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="registry-server" Dec 10 18:58:55 crc kubenswrapper[4894]: E1210 18:58:55.601832 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="extract-utilities" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601840 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="extract-utilities" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601975 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="89fbf7f8-3630-4587-95da-e8bf12127a64" containerName="route-controller-manager" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.601996 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe4fbd5-317b-49fd-9c30-82860839a273" containerName="controller-manager" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.602007 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b53e16f-8f52-48f3-bb9c-c018d2da6c48" containerName="pruner" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.602018 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2badda9c-032f-49d5-9ed1-8d8768e7c1ef" containerName="registry-server" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.602440 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.603875 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g"] Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.604007 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.604301 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.604338 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.604373 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.604867 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.604920 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.605932 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.607421 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.607422 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.607627 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.607640 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.607953 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.613087 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.613522 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.618075 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p"] Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.623262 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g"] Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bgh9\" (UniqueName: \"kubernetes.io/projected/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-kube-api-access-5bgh9\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704306 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-config\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704351 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-serving-cert\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704370 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-config\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704393 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-serving-cert\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704583 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-client-ca\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704677 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-client-ca\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704731 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-proxy-ca-bundles\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.704796 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g624q\" (UniqueName: \"kubernetes.io/projected/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-kube-api-access-g624q\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.805797 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bgh9\" (UniqueName: \"kubernetes.io/projected/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-kube-api-access-5bgh9\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.805943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-config\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-serving-cert\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807116 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-config\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807164 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-config\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-serving-cert\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807324 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-client-ca\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-client-ca\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807479 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-proxy-ca-bundles\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.807559 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g624q\" (UniqueName: \"kubernetes.io/projected/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-kube-api-access-g624q\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.808085 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-client-ca\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.808219 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-client-ca\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.808799 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-proxy-ca-bundles\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.809599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-config\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.811649 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-serving-cert\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.814730 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-serving-cert\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.826915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g624q\" (UniqueName: \"kubernetes.io/projected/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-kube-api-access-g624q\") pod \"controller-manager-54f9c4cc6d-5lg8g\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.827602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bgh9\" (UniqueName: \"kubernetes.io/projected/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-kube-api-access-5bgh9\") pod \"route-controller-manager-86dfb68df6-jmd6p\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.841773 4894 generic.go:334] "Generic (PLEG): container finished" podID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerID="9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2" exitCode=0 Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.841864 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerDied","Data":"9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2"} Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.929884 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:55 crc kubenswrapper[4894]: I1210 18:58:55.944099 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.154175 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p"] Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.431011 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g"] Dec 10 18:58:56 crc kubenswrapper[4894]: W1210 18:58:56.466525 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b5cd97_3367_4d7c_a1ef_67628e9ac9f4.slice/crio-4cd0bf0d0a4b45a45de8540b0e1e40a7d15110979f7bc3d8c84d6cf01cdc7b79 WatchSource:0}: Error finding container 4cd0bf0d0a4b45a45de8540b0e1e40a7d15110979f7bc3d8c84d6cf01cdc7b79: Status 404 returned error can't find the container with id 4cd0bf0d0a4b45a45de8540b0e1e40a7d15110979f7bc3d8c84d6cf01cdc7b79 Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.851756 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7tz9" event={"ID":"3a0cf23c-9387-4ed5-8f07-95b54c449786","Type":"ContainerStarted","Data":"3687ebcbd5f373a99944860e42e69ed7fee574127633972af3a2adf06843ccb5"} Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.854431 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mb95" event={"ID":"91c2b024-f9b4-4d85-bc5d-ad463ae51871","Type":"ContainerStarted","Data":"6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc"} Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.856480 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerID="95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.856530 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2zg" event={"ID":"2a2f64bf-a1f8-4ad3-9dd3-970afc101879","Type":"ContainerDied","Data":"95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7"} Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.858776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" event={"ID":"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4","Type":"ContainerStarted","Data":"4cd0bf0d0a4b45a45de8540b0e1e40a7d15110979f7bc3d8c84d6cf01cdc7b79"} Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.860062 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" event={"ID":"2d099a67-c165-4a14-a7f6-f6b70dae3b5a","Type":"ContainerStarted","Data":"084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6"} Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.860092 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" event={"ID":"2d099a67-c165-4a14-a7f6-f6b70dae3b5a","Type":"ContainerStarted","Data":"099afa9084256964a785de1725a8a0cefe3a87e77e8da09df31f70c366cfba09"} Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.860315 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.874530 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m7tz9" podStartSLOduration=4.798423639 podStartE2EDuration="59.874509618s" podCreationTimestamp="2025-12-10 18:57:57 +0000 UTC" firstStartedPulling="2025-12-10 18:58:00.128184605 +0000 UTC m=+162.923032397" lastFinishedPulling="2025-12-10 18:58:55.204270584 +0000 UTC m=+217.999118376" observedRunningTime="2025-12-10 18:58:56.872744918 +0000 UTC m=+219.667592710" watchObservedRunningTime="2025-12-10 18:58:56.874509618 +0000 UTC m=+219.669357410" Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.916358 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" podStartSLOduration=3.916332957 podStartE2EDuration="3.916332957s" podCreationTimestamp="2025-12-10 18:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:56.904904825 +0000 UTC m=+219.699752617" watchObservedRunningTime="2025-12-10 18:58:56.916332957 +0000 UTC m=+219.711180739" Dec 10 18:58:56 crc kubenswrapper[4894]: I1210 18:58:56.932597 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9mb95" podStartSLOduration=3.36538518 podStartE2EDuration="57.932579835s" podCreationTimestamp="2025-12-10 18:57:59 +0000 UTC" firstStartedPulling="2025-12-10 18:58:01.353212731 +0000 UTC m=+164.148060523" lastFinishedPulling="2025-12-10 18:58:55.920407386 +0000 UTC m=+218.715255178" observedRunningTime="2025-12-10 18:58:56.923987083 +0000 UTC m=+219.718834885" watchObservedRunningTime="2025-12-10 18:58:56.932579835 +0000 UTC m=+219.727427627" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.084712 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.866601 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" event={"ID":"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4","Type":"ContainerStarted","Data":"5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8"} Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.867027 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.872565 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.886201 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" podStartSLOduration=4.886184513 podStartE2EDuration="4.886184513s" podCreationTimestamp="2025-12-10 18:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:57.883560749 +0000 UTC m=+220.678408541" watchObservedRunningTime="2025-12-10 18:58:57.886184513 +0000 UTC m=+220.681032305" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.916054 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.916100 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:58:57 crc kubenswrapper[4894]: I1210 18:58:57.987479 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.049125 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.049462 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.082923 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.084059 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.084112 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.128244 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:58:58 crc kubenswrapper[4894]: I1210 18:58:58.918448 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 18:58:59 crc kubenswrapper[4894]: I1210 18:58:59.727936 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:58:59 crc kubenswrapper[4894]: I1210 18:58:59.728028 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:58:59 crc kubenswrapper[4894]: I1210 18:58:59.767654 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:58:59 crc kubenswrapper[4894]: I1210 18:58:59.878410 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerStarted","Data":"76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68"} Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.107667 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.126744 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kt5vd" podStartSLOduration=13.434690916 podStartE2EDuration="1m9.126726217s" podCreationTimestamp="2025-12-10 18:57:59 +0000 UTC" firstStartedPulling="2025-12-10 18:58:01.35603746 +0000 UTC m=+164.150885252" lastFinishedPulling="2025-12-10 18:58:57.048072751 +0000 UTC m=+219.842920553" observedRunningTime="2025-12-10 18:59:00.900399972 +0000 UTC m=+223.695247764" watchObservedRunningTime="2025-12-10 18:59:08.126726217 +0000 UTC m=+230.921573999" Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.134606 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.153167 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m7tz9"] Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.937197 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2zg" event={"ID":"2a2f64bf-a1f8-4ad3-9dd3-970afc101879","Type":"ContainerStarted","Data":"0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b"} Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.937440 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m7tz9" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="registry-server" containerID="cri-o://3687ebcbd5f373a99944860e42e69ed7fee574127633972af3a2adf06843ccb5" gracePeriod=2 Dec 10 18:59:08 crc kubenswrapper[4894]: I1210 18:59:08.967246 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vr2zg" podStartSLOduration=5.5410425629999995 podStartE2EDuration="1m11.967225495s" podCreationTimestamp="2025-12-10 18:57:57 +0000 UTC" firstStartedPulling="2025-12-10 18:58:00.234181898 +0000 UTC m=+163.029029690" lastFinishedPulling="2025-12-10 18:59:06.66036483 +0000 UTC m=+229.455212622" observedRunningTime="2025-12-10 18:59:08.965788025 +0000 UTC m=+231.760635827" watchObservedRunningTime="2025-12-10 18:59:08.967225495 +0000 UTC m=+231.762073297" Dec 10 18:59:09 crc kubenswrapper[4894]: I1210 18:59:09.746506 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dkr5p"] Dec 10 18:59:09 crc kubenswrapper[4894]: I1210 18:59:09.746767 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dkr5p" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="registry-server" containerID="cri-o://db0aa1826da1f98708f7e6aee7db016fe6fcb7cc710455613cf3bc808c1ae3de" gracePeriod=2 Dec 10 18:59:09 crc kubenswrapper[4894]: I1210 18:59:09.790138 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 18:59:10 crc kubenswrapper[4894]: I1210 18:59:10.207517 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:59:10 crc kubenswrapper[4894]: I1210 18:59:10.208169 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:59:10 crc kubenswrapper[4894]: I1210 18:59:10.271643 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.001622 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.957542 4894 generic.go:334] "Generic (PLEG): container finished" podID="dada86bd-f99d-4566-98eb-33194725df47" containerID="db0aa1826da1f98708f7e6aee7db016fe6fcb7cc710455613cf3bc808c1ae3de" exitCode=0 Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.957613 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerDied","Data":"db0aa1826da1f98708f7e6aee7db016fe6fcb7cc710455613cf3bc808c1ae3de"} Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.957904 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dkr5p" event={"ID":"dada86bd-f99d-4566-98eb-33194725df47","Type":"ContainerDied","Data":"d3070f617cf8cef40c9c2b50547f3db6ad3264ad1acb6432eeceb016b41fcd82"} Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.957947 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3070f617cf8cef40c9c2b50547f3db6ad3264ad1acb6432eeceb016b41fcd82" Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.960528 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.961715 4894 generic.go:334] "Generic (PLEG): container finished" podID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerID="3687ebcbd5f373a99944860e42e69ed7fee574127633972af3a2adf06843ccb5" exitCode=0 Dec 10 18:59:11 crc kubenswrapper[4894]: I1210 18:59:11.962460 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7tz9" event={"ID":"3a0cf23c-9387-4ed5-8f07-95b54c449786","Type":"ContainerDied","Data":"3687ebcbd5f373a99944860e42e69ed7fee574127633972af3a2adf06843ccb5"} Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.019549 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.039894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-utilities\") pod \"dada86bd-f99d-4566-98eb-33194725df47\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.040013 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-catalog-content\") pod \"dada86bd-f99d-4566-98eb-33194725df47\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.040094 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljt58\" (UniqueName: \"kubernetes.io/projected/dada86bd-f99d-4566-98eb-33194725df47-kube-api-access-ljt58\") pod \"dada86bd-f99d-4566-98eb-33194725df47\" (UID: \"dada86bd-f99d-4566-98eb-33194725df47\") " Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.040914 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-utilities" (OuterVolumeSpecName: "utilities") pod "dada86bd-f99d-4566-98eb-33194725df47" (UID: "dada86bd-f99d-4566-98eb-33194725df47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.047048 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dada86bd-f99d-4566-98eb-33194725df47-kube-api-access-ljt58" (OuterVolumeSpecName: "kube-api-access-ljt58") pod "dada86bd-f99d-4566-98eb-33194725df47" (UID: "dada86bd-f99d-4566-98eb-33194725df47"). InnerVolumeSpecName "kube-api-access-ljt58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.085340 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dada86bd-f99d-4566-98eb-33194725df47" (UID: "dada86bd-f99d-4566-98eb-33194725df47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.141791 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqxz4\" (UniqueName: \"kubernetes.io/projected/3a0cf23c-9387-4ed5-8f07-95b54c449786-kube-api-access-bqxz4\") pod \"3a0cf23c-9387-4ed5-8f07-95b54c449786\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.141928 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-utilities\") pod \"3a0cf23c-9387-4ed5-8f07-95b54c449786\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.141978 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-catalog-content\") pod \"3a0cf23c-9387-4ed5-8f07-95b54c449786\" (UID: \"3a0cf23c-9387-4ed5-8f07-95b54c449786\") " Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.142246 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.142265 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dada86bd-f99d-4566-98eb-33194725df47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.142279 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljt58\" (UniqueName: \"kubernetes.io/projected/dada86bd-f99d-4566-98eb-33194725df47-kube-api-access-ljt58\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.143820 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-utilities" (OuterVolumeSpecName: "utilities") pod "3a0cf23c-9387-4ed5-8f07-95b54c449786" (UID: "3a0cf23c-9387-4ed5-8f07-95b54c449786"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.146417 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a0cf23c-9387-4ed5-8f07-95b54c449786-kube-api-access-bqxz4" (OuterVolumeSpecName: "kube-api-access-bqxz4") pod "3a0cf23c-9387-4ed5-8f07-95b54c449786" (UID: "3a0cf23c-9387-4ed5-8f07-95b54c449786"). InnerVolumeSpecName "kube-api-access-bqxz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.194061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a0cf23c-9387-4ed5-8f07-95b54c449786" (UID: "3a0cf23c-9387-4ed5-8f07-95b54c449786"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.243287 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.243320 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0cf23c-9387-4ed5-8f07-95b54c449786-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.243331 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqxz4\" (UniqueName: \"kubernetes.io/projected/3a0cf23c-9387-4ed5-8f07-95b54c449786-kube-api-access-bqxz4\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.969153 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m7tz9" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.969138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m7tz9" event={"ID":"3a0cf23c-9387-4ed5-8f07-95b54c449786","Type":"ContainerDied","Data":"ff0986d43253fe4b7a0574e3b11f952e174c2ad751e4f2d5f0f4db4d52c0da1b"} Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.969179 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dkr5p" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.969806 4894 scope.go:117] "RemoveContainer" containerID="3687ebcbd5f373a99944860e42e69ed7fee574127633972af3a2adf06843ccb5" Dec 10 18:59:12 crc kubenswrapper[4894]: I1210 18:59:12.990175 4894 scope.go:117] "RemoveContainer" containerID="0fbc5bd9d991bc1189bd4818a253c28450d111f9d232bdfe32ba06136a6c218a" Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.019161 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m7tz9"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.019502 4894 scope.go:117] "RemoveContainer" containerID="424d110b73b06af3a13f063dac8cd6004c0333e9a1085e49be0e72c0d5a689a5" Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.022412 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m7tz9"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.034559 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dkr5p"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.040804 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dkr5p"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.547073 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" path="/var/lib/kubelet/pods/3a0cf23c-9387-4ed5-8f07-95b54c449786/volumes" Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.547647 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dada86bd-f99d-4566-98eb-33194725df47" path="/var/lib/kubelet/pods/dada86bd-f99d-4566-98eb-33194725df47/volumes" Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.845450 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m2s4d"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.906782 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.907047 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" podUID="73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" containerName="controller-manager" containerID="cri-o://5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8" gracePeriod=30 Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.950409 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p"] Dec 10 18:59:13 crc kubenswrapper[4894]: I1210 18:59:13.950871 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" podUID="2d099a67-c165-4a14-a7f6-f6b70dae3b5a" containerName="route-controller-manager" containerID="cri-o://084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6" gracePeriod=30 Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.342465 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kt5vd"] Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.342709 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kt5vd" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="registry-server" containerID="cri-o://76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68" gracePeriod=2 Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.497711 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.572403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-config\") pod \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.572440 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-serving-cert\") pod \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.572515 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-client-ca\") pod \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.572548 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bgh9\" (UniqueName: \"kubernetes.io/projected/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-kube-api-access-5bgh9\") pod \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\" (UID: \"2d099a67-c165-4a14-a7f6-f6b70dae3b5a\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.573295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-client-ca" (OuterVolumeSpecName: "client-ca") pod "2d099a67-c165-4a14-a7f6-f6b70dae3b5a" (UID: "2d099a67-c165-4a14-a7f6-f6b70dae3b5a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.573346 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-config" (OuterVolumeSpecName: "config") pod "2d099a67-c165-4a14-a7f6-f6b70dae3b5a" (UID: "2d099a67-c165-4a14-a7f6-f6b70dae3b5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.577759 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-kube-api-access-5bgh9" (OuterVolumeSpecName: "kube-api-access-5bgh9") pod "2d099a67-c165-4a14-a7f6-f6b70dae3b5a" (UID: "2d099a67-c165-4a14-a7f6-f6b70dae3b5a"). InnerVolumeSpecName "kube-api-access-5bgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.578219 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2d099a67-c165-4a14-a7f6-f6b70dae3b5a" (UID: "2d099a67-c165-4a14-a7f6-f6b70dae3b5a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.580491 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.673860 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-config\") pod \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.673940 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-client-ca\") pod \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674036 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-serving-cert\") pod \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674055 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-proxy-ca-bundles\") pod \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674090 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g624q\" (UniqueName: \"kubernetes.io/projected/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-kube-api-access-g624q\") pod \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\" (UID: \"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674263 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674275 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674283 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674292 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bgh9\" (UniqueName: \"kubernetes.io/projected/2d099a67-c165-4a14-a7f6-f6b70dae3b5a-kube-api-access-5bgh9\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674770 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-config" (OuterVolumeSpecName: "config") pod "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" (UID: "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674876 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" (UID: "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.674903 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-client-ca" (OuterVolumeSpecName: "client-ca") pod "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" (UID: "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.679959 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-kube-api-access-g624q" (OuterVolumeSpecName: "kube-api-access-g624q") pod "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" (UID: "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4"). InnerVolumeSpecName "kube-api-access-g624q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.680979 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" (UID: "73b5cd97-3367-4d7c-a1ef-67628e9ac9f4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.717264 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.775621 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-utilities\") pod \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.775666 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvwcp\" (UniqueName: \"kubernetes.io/projected/f8cb70e6-c927-4dea-8184-ae0a5343ddec-kube-api-access-jvwcp\") pod \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.775785 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-catalog-content\") pod \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\" (UID: \"f8cb70e6-c927-4dea-8184-ae0a5343ddec\") " Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.776605 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-utilities" (OuterVolumeSpecName: "utilities") pod "f8cb70e6-c927-4dea-8184-ae0a5343ddec" (UID: "f8cb70e6-c927-4dea-8184-ae0a5343ddec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.778425 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8cb70e6-c927-4dea-8184-ae0a5343ddec-kube-api-access-jvwcp" (OuterVolumeSpecName: "kube-api-access-jvwcp") pod "f8cb70e6-c927-4dea-8184-ae0a5343ddec" (UID: "f8cb70e6-c927-4dea-8184-ae0a5343ddec"). InnerVolumeSpecName "kube-api-access-jvwcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791716 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791747 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791759 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g624q\" (UniqueName: \"kubernetes.io/projected/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-kube-api-access-g624q\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791768 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791776 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791784 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvwcp\" (UniqueName: \"kubernetes.io/projected/f8cb70e6-c927-4dea-8184-ae0a5343ddec-kube-api-access-jvwcp\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.791794 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.802946 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8cb70e6-c927-4dea-8184-ae0a5343ddec" (UID: "f8cb70e6-c927-4dea-8184-ae0a5343ddec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.893397 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cb70e6-c927-4dea-8184-ae0a5343ddec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.981681 4894 generic.go:334] "Generic (PLEG): container finished" podID="2d099a67-c165-4a14-a7f6-f6b70dae3b5a" containerID="084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6" exitCode=0 Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.981737 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.981743 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" event={"ID":"2d099a67-c165-4a14-a7f6-f6b70dae3b5a","Type":"ContainerDied","Data":"084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6"} Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.981866 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p" event={"ID":"2d099a67-c165-4a14-a7f6-f6b70dae3b5a","Type":"ContainerDied","Data":"099afa9084256964a785de1725a8a0cefe3a87e77e8da09df31f70c366cfba09"} Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.981888 4894 scope.go:117] "RemoveContainer" containerID="084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.984965 4894 generic.go:334] "Generic (PLEG): container finished" podID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerID="76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68" exitCode=0 Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.985015 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerDied","Data":"76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68"} Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.985031 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kt5vd" event={"ID":"f8cb70e6-c927-4dea-8184-ae0a5343ddec","Type":"ContainerDied","Data":"35ea39d0807d6860b7da8143caba2f031501699749621ddd091d794fbf637912"} Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.985091 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kt5vd" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.986494 4894 generic.go:334] "Generic (PLEG): container finished" podID="73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" containerID="5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8" exitCode=0 Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.986555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" event={"ID":"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4","Type":"ContainerDied","Data":"5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8"} Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.986571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" event={"ID":"73b5cd97-3367-4d7c-a1ef-67628e9ac9f4","Type":"ContainerDied","Data":"4cd0bf0d0a4b45a45de8540b0e1e40a7d15110979f7bc3d8c84d6cf01cdc7b79"} Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.986620 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g" Dec 10 18:59:14 crc kubenswrapper[4894]: I1210 18:59:14.998911 4894 scope.go:117] "RemoveContainer" containerID="084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:14.999387 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6\": container with ID starting with 084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6 not found: ID does not exist" containerID="084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:14.999432 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6"} err="failed to get container status \"084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6\": rpc error: code = NotFound desc = could not find container \"084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6\": container with ID starting with 084c09f14e0695b63f02b34ebdf17e81df0b86a8d1739170a8ce650ace4f35a6 not found: ID does not exist" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:14.999462 4894 scope.go:117] "RemoveContainer" containerID="76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.028183 4894 scope.go:117] "RemoveContainer" containerID="9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.033004 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.037587 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dfb68df6-jmd6p"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.042031 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.045487 4894 scope.go:117] "RemoveContainer" containerID="eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.053922 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-54f9c4cc6d-5lg8g"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.059362 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kt5vd"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.064604 4894 scope.go:117] "RemoveContainer" containerID="76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.066961 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kt5vd"] Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.071270 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68\": container with ID starting with 76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68 not found: ID does not exist" containerID="76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.071396 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68"} err="failed to get container status \"76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68\": rpc error: code = NotFound desc = could not find container \"76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68\": container with ID starting with 76f22d7e63b1cdeb72f4717f80ab4ddc4131aad147fa67cfc161cc72b4b9da68 not found: ID does not exist" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.071516 4894 scope.go:117] "RemoveContainer" containerID="9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.072080 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2\": container with ID starting with 9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2 not found: ID does not exist" containerID="9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.072191 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2"} err="failed to get container status \"9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2\": rpc error: code = NotFound desc = could not find container \"9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2\": container with ID starting with 9cb7a48c35c54a8568ecb96965831aa19e4738d192c1ec1d45c03f02babad7b2 not found: ID does not exist" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.072273 4894 scope.go:117] "RemoveContainer" containerID="eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.072534 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec\": container with ID starting with eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec not found: ID does not exist" containerID="eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.072629 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec"} err="failed to get container status \"eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec\": rpc error: code = NotFound desc = could not find container \"eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec\": container with ID starting with eba2813e924e0073c6e8084e80655c00157c614325197478b124777cf7ba46ec not found: ID does not exist" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.072713 4894 scope.go:117] "RemoveContainer" containerID="5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.093063 4894 scope.go:117] "RemoveContainer" containerID="5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.093521 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8\": container with ID starting with 5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8 not found: ID does not exist" containerID="5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.093632 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8"} err="failed to get container status \"5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8\": rpc error: code = NotFound desc = could not find container \"5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8\": container with ID starting with 5ad41d8d80ba0487191f26a03e32ffc1a390b95322d51cde32f67897dbdb55e8 not found: ID does not exist" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.548204 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d099a67-c165-4a14-a7f6-f6b70dae3b5a" path="/var/lib/kubelet/pods/2d099a67-c165-4a14-a7f6-f6b70dae3b5a/volumes" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.548883 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" path="/var/lib/kubelet/pods/73b5cd97-3367-4d7c-a1ef-67628e9ac9f4/volumes" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.549492 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" path="/var/lib/kubelet/pods/f8cb70e6-c927-4dea-8184-ae0a5343ddec/volumes" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618249 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r"] Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618522 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" containerName="controller-manager" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618543 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" containerName="controller-manager" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618558 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618568 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618579 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="extract-utilities" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618586 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="extract-utilities" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618595 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="extract-content" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618601 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="extract-content" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618608 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="extract-content" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618614 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="extract-content" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618623 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="extract-utilities" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618640 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="extract-utilities" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618654 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="extract-utilities" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618661 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="extract-utilities" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618670 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="extract-content" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618677 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="extract-content" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618689 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618697 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618710 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618719 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: E1210 18:59:15.618728 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d099a67-c165-4a14-a7f6-f6b70dae3b5a" containerName="route-controller-manager" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618741 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d099a67-c165-4a14-a7f6-f6b70dae3b5a" containerName="route-controller-manager" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618866 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d099a67-c165-4a14-a7f6-f6b70dae3b5a" containerName="route-controller-manager" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618882 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0cf23c-9387-4ed5-8f07-95b54c449786" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618892 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b5cd97-3367-4d7c-a1ef-67628e9ac9f4" containerName="controller-manager" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618905 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="dada86bd-f99d-4566-98eb-33194725df47" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.618916 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8cb70e6-c927-4dea-8184-ae0a5343ddec" containerName="registry-server" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.619359 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.621597 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-bff6fd464-fc2t4"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.621761 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.622172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.622317 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.622327 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.622838 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.623041 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.623240 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.625543 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.625569 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.625732 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.625778 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.626109 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.627933 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.634767 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.634871 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.651420 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bff6fd464-fc2t4"] Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701581 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-config\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701638 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0788ca5-497a-40a2-9335-acdb4b97cfa4-config\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701764 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-proxy-ca-bundles\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701877 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4k9m\" (UniqueName: \"kubernetes.io/projected/47636a50-e92e-42f1-8d5c-2c614d4e942e-kube-api-access-r4k9m\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p48sp\" (UniqueName: \"kubernetes.io/projected/f0788ca5-497a-40a2-9335-acdb4b97cfa4-kube-api-access-p48sp\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47636a50-e92e-42f1-8d5c-2c614d4e942e-serving-cert\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.701984 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-client-ca\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.702063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0788ca5-497a-40a2-9335-acdb4b97cfa4-serving-cert\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.702105 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0788ca5-497a-40a2-9335-acdb4b97cfa4-client-ca\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.803682 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0788ca5-497a-40a2-9335-acdb4b97cfa4-client-ca\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.803843 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-config\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.803958 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0788ca5-497a-40a2-9335-acdb4b97cfa4-config\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.804032 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-proxy-ca-bundles\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.804081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4k9m\" (UniqueName: \"kubernetes.io/projected/47636a50-e92e-42f1-8d5c-2c614d4e942e-kube-api-access-r4k9m\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.804117 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p48sp\" (UniqueName: \"kubernetes.io/projected/f0788ca5-497a-40a2-9335-acdb4b97cfa4-kube-api-access-p48sp\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.804152 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-client-ca\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.804185 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47636a50-e92e-42f1-8d5c-2c614d4e942e-serving-cert\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.804228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0788ca5-497a-40a2-9335-acdb4b97cfa4-serving-cert\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.805213 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-client-ca\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.805953 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0788ca5-497a-40a2-9335-acdb4b97cfa4-client-ca\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.806117 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0788ca5-497a-40a2-9335-acdb4b97cfa4-config\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.807679 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-config\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.810709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47636a50-e92e-42f1-8d5c-2c614d4e942e-proxy-ca-bundles\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.817649 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0788ca5-497a-40a2-9335-acdb4b97cfa4-serving-cert\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.822034 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47636a50-e92e-42f1-8d5c-2c614d4e942e-serving-cert\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.822800 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4k9m\" (UniqueName: \"kubernetes.io/projected/47636a50-e92e-42f1-8d5c-2c614d4e942e-kube-api-access-r4k9m\") pod \"controller-manager-bff6fd464-fc2t4\" (UID: \"47636a50-e92e-42f1-8d5c-2c614d4e942e\") " pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.825226 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p48sp\" (UniqueName: \"kubernetes.io/projected/f0788ca5-497a-40a2-9335-acdb4b97cfa4-kube-api-access-p48sp\") pod \"route-controller-manager-656977689d-kvd5r\" (UID: \"f0788ca5-497a-40a2-9335-acdb4b97cfa4\") " pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.938535 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:15 crc kubenswrapper[4894]: I1210 18:59:15.953091 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:16 crc kubenswrapper[4894]: I1210 18:59:16.201230 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r"] Dec 10 18:59:16 crc kubenswrapper[4894]: I1210 18:59:16.254366 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bff6fd464-fc2t4"] Dec 10 18:59:16 crc kubenswrapper[4894]: W1210 18:59:16.265074 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47636a50_e92e_42f1_8d5c_2c614d4e942e.slice/crio-fab44c9906c1ef8370a666fb7ed4f5836366ac8308e9461d97b5d4ba9a98c6b7 WatchSource:0}: Error finding container fab44c9906c1ef8370a666fb7ed4f5836366ac8308e9461d97b5d4ba9a98c6b7: Status 404 returned error can't find the container with id fab44c9906c1ef8370a666fb7ed4f5836366ac8308e9461d97b5d4ba9a98c6b7 Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.006044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" event={"ID":"47636a50-e92e-42f1-8d5c-2c614d4e942e","Type":"ContainerStarted","Data":"a34887a68119fdac891c292c66ea9543d971e2dca264133f96db1df6a9bb9158"} Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.006097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" event={"ID":"47636a50-e92e-42f1-8d5c-2c614d4e942e","Type":"ContainerStarted","Data":"fab44c9906c1ef8370a666fb7ed4f5836366ac8308e9461d97b5d4ba9a98c6b7"} Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.006116 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.009097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" event={"ID":"f0788ca5-497a-40a2-9335-acdb4b97cfa4","Type":"ContainerStarted","Data":"ea4479e8a28c574917eec4d3c2b32d9fcb8df32e4f337fce1bc6ce0010ee7a38"} Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.009153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" event={"ID":"f0788ca5-497a-40a2-9335-acdb4b97cfa4","Type":"ContainerStarted","Data":"b6e68f7988cf2b0731137c91efc070d25b7a59636b513080680f3990cf368350"} Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.009268 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.010778 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.013771 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.030171 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-bff6fd464-fc2t4" podStartSLOduration=4.030154358 podStartE2EDuration="4.030154358s" podCreationTimestamp="2025-12-10 18:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:17.025877367 +0000 UTC m=+239.820725159" watchObservedRunningTime="2025-12-10 18:59:17.030154358 +0000 UTC m=+239.825002150" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.070454 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-656977689d-kvd5r" podStartSLOduration=4.070438564 podStartE2EDuration="4.070438564s" podCreationTimestamp="2025-12-10 18:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:17.06817013 +0000 UTC m=+239.863017922" watchObservedRunningTime="2025-12-10 18:59:17.070438564 +0000 UTC m=+239.865286356" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.539364 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.542242 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:59:17 crc kubenswrapper[4894]: I1210 18:59:17.612073 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:59:18 crc kubenswrapper[4894]: I1210 18:59:18.065065 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.189816 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.191818 4894 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.191946 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.191979 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192172 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d" gracePeriod=15 Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192218 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09" gracePeriod=15 Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.192396 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192497 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.192565 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192632 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.192718 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192784 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.192864 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192950 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.193580 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.193641 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.193702 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.193757 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.193826 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.193914 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192350 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c" gracePeriod=15 Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192292 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4" gracePeriod=15 Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.192425 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7" gracePeriod=15 Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.194336 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.194504 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.194620 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.194676 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.194732 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.194786 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.204981 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.233968 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.286437 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.286734 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.286889 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.287005 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.287143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.287261 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.287380 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.287486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388246 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388615 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388637 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388671 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388688 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388711 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388728 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388799 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388839 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388871 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388892 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388912 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.388949 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: I1210 18:59:22.527714 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:22 crc kubenswrapper[4894]: W1210 18:59:22.550504 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-e37b4d2e9633e59b420ce905745f67261b07c66916783a3c53664d4c0581c85c WatchSource:0}: Error finding container e37b4d2e9633e59b420ce905745f67261b07c66916783a3c53664d4c0581c85c: Status 404 returned error can't find the container with id e37b4d2e9633e59b420ce905745f67261b07c66916783a3c53664d4c0581c85c Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.554069 4894 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.201:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" volumeName="registry-storage" Dec 10 18:59:22 crc kubenswrapper[4894]: E1210 18:59:22.554033 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.201:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187fefba56e3cc82 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:59:22.553248898 +0000 UTC m=+245.348096730,LastTimestamp:2025-12-10 18:59:22.553248898 +0000 UTC m=+245.348096730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.070343 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.073337 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.074690 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09" exitCode=0 Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.074718 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c" exitCode=0 Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.074729 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7" exitCode=0 Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.074740 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4" exitCode=2 Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.074814 4894 scope.go:117] "RemoveContainer" containerID="b2bc16948c4c9226be0bb40a95484ad220821e2494447b4720b7697b83bce066" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.077568 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" containerID="a8982172dfc36ee4653ff5b0182fd7d5699f08fbc017842c7684cdd1f1d83ffe" exitCode=0 Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.077655 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e5bc823d-aae9-4d4e-b6cd-095590924c6c","Type":"ContainerDied","Data":"a8982172dfc36ee4653ff5b0182fd7d5699f08fbc017842c7684cdd1f1d83ffe"} Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.078571 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.079464 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.080408 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"21e93b9b72eb748ce7e5d3ec955009922ec5e52c59d77dd90392a737369e18b2"} Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.080468 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e37b4d2e9633e59b420ce905745f67261b07c66916783a3c53664d4c0581c85c"} Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.081003 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4894]: I1210 18:59:23.081449 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.091464 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.564087 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.565182 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.565593 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.566725 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.567537 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.567795 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.567990 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.568250 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617472 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kube-api-access\") pod \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617534 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617573 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617643 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kubelet-dir\") pod \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617670 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-var-lock\") pod \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\" (UID: \"e5bc823d-aae9-4d4e-b6cd-095590924c6c\") " Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617663 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617687 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617694 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617747 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e5bc823d-aae9-4d4e-b6cd-095590924c6c" (UID: "e5bc823d-aae9-4d4e-b6cd-095590924c6c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.617798 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-var-lock" (OuterVolumeSpecName: "var-lock") pod "e5bc823d-aae9-4d4e-b6cd-095590924c6c" (UID: "e5bc823d-aae9-4d4e-b6cd-095590924c6c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.618763 4894 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.618779 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.618788 4894 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e5bc823d-aae9-4d4e-b6cd-095590924c6c-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.618796 4894 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.618805 4894 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.625832 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e5bc823d-aae9-4d4e-b6cd-095590924c6c" (UID: "e5bc823d-aae9-4d4e-b6cd-095590924c6c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:24 crc kubenswrapper[4894]: I1210 18:59:24.720164 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5bc823d-aae9-4d4e-b6cd-095590924c6c-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.106095 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.107587 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d" exitCode=0 Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.107799 4894 scope.go:117] "RemoveContainer" containerID="3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.107925 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.110397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e5bc823d-aae9-4d4e-b6cd-095590924c6c","Type":"ContainerDied","Data":"e0c64a1098e0246964ab6c40ae84f25d623d32af8421bba8ec4c03e489e08da2"} Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.110445 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0c64a1098e0246964ab6c40ae84f25d623d32af8421bba8ec4c03e489e08da2" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.110458 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.133684 4894 scope.go:117] "RemoveContainer" containerID="83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.137353 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.137780 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.139046 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.148551 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.149208 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.149837 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.151815 4894 scope.go:117] "RemoveContainer" containerID="b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.172287 4894 scope.go:117] "RemoveContainer" containerID="5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.193223 4894 scope.go:117] "RemoveContainer" containerID="ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.216219 4894 scope.go:117] "RemoveContainer" containerID="cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.241180 4894 scope.go:117] "RemoveContainer" containerID="3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09" Dec 10 18:59:25 crc kubenswrapper[4894]: E1210 18:59:25.241820 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\": container with ID starting with 3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09 not found: ID does not exist" containerID="3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.243033 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09"} err="failed to get container status \"3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\": rpc error: code = NotFound desc = could not find container \"3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09\": container with ID starting with 3575cf7c23a384d82c2a5b0d44d3458c4a14717968c33809388309f30d76fc09 not found: ID does not exist" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.243126 4894 scope.go:117] "RemoveContainer" containerID="83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c" Dec 10 18:59:25 crc kubenswrapper[4894]: E1210 18:59:25.244424 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\": container with ID starting with 83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c not found: ID does not exist" containerID="83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.244556 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c"} err="failed to get container status \"83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\": rpc error: code = NotFound desc = could not find container \"83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c\": container with ID starting with 83ab53a4bfe4b4767d40a8bdbee68b9ed4f2921ed3ce229cb7b44b37a1dfa22c not found: ID does not exist" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.244666 4894 scope.go:117] "RemoveContainer" containerID="b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7" Dec 10 18:59:25 crc kubenswrapper[4894]: E1210 18:59:25.245239 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\": container with ID starting with b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7 not found: ID does not exist" containerID="b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.245296 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7"} err="failed to get container status \"b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\": rpc error: code = NotFound desc = could not find container \"b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7\": container with ID starting with b3ec01a5ea17a640c40c3701d96d5bfe34ff57b0b5510726b39a6a26d81977d7 not found: ID does not exist" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.245335 4894 scope.go:117] "RemoveContainer" containerID="5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4" Dec 10 18:59:25 crc kubenswrapper[4894]: E1210 18:59:25.245800 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\": container with ID starting with 5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4 not found: ID does not exist" containerID="5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.245984 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4"} err="failed to get container status \"5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\": rpc error: code = NotFound desc = could not find container \"5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4\": container with ID starting with 5af6bab78f6c719ab652bf9c45a357e8ea603a60448b1bbfcd179f1ff5594db4 not found: ID does not exist" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.246107 4894 scope.go:117] "RemoveContainer" containerID="ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d" Dec 10 18:59:25 crc kubenswrapper[4894]: E1210 18:59:25.246946 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\": container with ID starting with ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d not found: ID does not exist" containerID="ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.246999 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d"} err="failed to get container status \"ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\": rpc error: code = NotFound desc = could not find container \"ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d\": container with ID starting with ce71ba8f2bb27afc75f6daf3428a390e7bb81e0006b0565c537ecbd2e78c056d not found: ID does not exist" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.247030 4894 scope.go:117] "RemoveContainer" containerID="cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b" Dec 10 18:59:25 crc kubenswrapper[4894]: E1210 18:59:25.247417 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\": container with ID starting with cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b not found: ID does not exist" containerID="cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.247561 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b"} err="failed to get container status \"cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\": rpc error: code = NotFound desc = could not find container \"cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b\": container with ID starting with cf5eb6aa58f14f1dbd4578e79cfad21fc057ccdf90bebb5022d94972b223f10b not found: ID does not exist" Dec 10 18:59:25 crc kubenswrapper[4894]: I1210 18:59:25.551178 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.623437 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.623979 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.624260 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.624510 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.624724 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4894]: I1210 18:59:26.624747 4894 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.624982 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="200ms" Dec 10 18:59:26 crc kubenswrapper[4894]: E1210 18:59:26.825953 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="400ms" Dec 10 18:59:27 crc kubenswrapper[4894]: E1210 18:59:27.227234 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="800ms" Dec 10 18:59:27 crc kubenswrapper[4894]: I1210 18:59:27.547014 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:27 crc kubenswrapper[4894]: I1210 18:59:27.547577 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:28 crc kubenswrapper[4894]: E1210 18:59:28.028240 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="1.6s" Dec 10 18:59:28 crc kubenswrapper[4894]: E1210 18:59:28.627419 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.201:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187fefba56e3cc82 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:59:22.553248898 +0000 UTC m=+245.348096730,LastTimestamp:2025-12-10 18:59:22.553248898 +0000 UTC m=+245.348096730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 18:59:29 crc kubenswrapper[4894]: E1210 18:59:29.629635 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="3.2s" Dec 10 18:59:32 crc kubenswrapper[4894]: E1210 18:59:32.830607 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="6.4s" Dec 10 18:59:33 crc kubenswrapper[4894]: I1210 18:59:33.540670 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:33 crc kubenswrapper[4894]: I1210 18:59:33.541717 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:33 crc kubenswrapper[4894]: I1210 18:59:33.542408 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:33 crc kubenswrapper[4894]: I1210 18:59:33.568010 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:33 crc kubenswrapper[4894]: I1210 18:59:33.568057 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:33 crc kubenswrapper[4894]: E1210 18:59:33.568546 4894 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:33 crc kubenswrapper[4894]: I1210 18:59:33.569276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:33 crc kubenswrapper[4894]: W1210 18:59:33.592462 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-36e444a2de813f5e762ae842fce0ea74dfc24b3a7e584a13b6253bddb35e9868 WatchSource:0}: Error finding container 36e444a2de813f5e762ae842fce0ea74dfc24b3a7e584a13b6253bddb35e9868: Status 404 returned error can't find the container with id 36e444a2de813f5e762ae842fce0ea74dfc24b3a7e584a13b6253bddb35e9868 Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.178645 4894 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="5287d980b767012f7dace490d6e2e045d4ea80395de166bdb995672bef16e306" exitCode=0 Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.178785 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"5287d980b767012f7dace490d6e2e045d4ea80395de166bdb995672bef16e306"} Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.179163 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"36e444a2de813f5e762ae842fce0ea74dfc24b3a7e584a13b6253bddb35e9868"} Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.179645 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.179675 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.180174 4894 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:34 crc kubenswrapper[4894]: E1210 18:59:34.180184 4894 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:34 crc kubenswrapper[4894]: I1210 18:59:34.180665 4894 status_manager.go:851] "Failed to get status for pod" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 18:59:35 crc kubenswrapper[4894]: I1210 18:59:35.187576 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"63cdc02f4b1aca2991e9b3e2f4b1140f1d637f8809796debb13ab43b734b637b"} Dec 10 18:59:35 crc kubenswrapper[4894]: I1210 18:59:35.187839 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"aebba86c1fff4b8911379200c6e15f763bec0673afb73f0822bc8a33109b7a27"} Dec 10 18:59:35 crc kubenswrapper[4894]: I1210 18:59:35.187862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a025f70968fae154e24c2fa7499838bf168c3991cd9366582c96c0594bc2b349"} Dec 10 18:59:36 crc kubenswrapper[4894]: I1210 18:59:36.195916 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f0f372d2855a39f023f6bf9d4001269f2e5ca1a8fed8ceb8acbcf2da732d1cf7"} Dec 10 18:59:36 crc kubenswrapper[4894]: I1210 18:59:36.195961 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"be6d62c4e50ddd96027191624c2e12219daa345260ea55450f92b561768392fe"} Dec 10 18:59:36 crc kubenswrapper[4894]: I1210 18:59:36.196094 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:36 crc kubenswrapper[4894]: I1210 18:59:36.196182 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:36 crc kubenswrapper[4894]: I1210 18:59:36.196206 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:37 crc kubenswrapper[4894]: I1210 18:59:37.206992 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 18:59:37 crc kubenswrapper[4894]: I1210 18:59:37.207901 4894 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369" exitCode=1 Dec 10 18:59:37 crc kubenswrapper[4894]: I1210 18:59:37.207953 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369"} Dec 10 18:59:37 crc kubenswrapper[4894]: I1210 18:59:37.208645 4894 scope.go:117] "RemoveContainer" containerID="a8e4b7df1081ad9a1de47d3d637a1bb38b7b0d85a09323ef5d32076c94424369" Dec 10 18:59:38 crc kubenswrapper[4894]: I1210 18:59:38.220370 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 18:59:38 crc kubenswrapper[4894]: I1210 18:59:38.220443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"154d40631e37b4644bbba2cc713572f0de1b624a9fb54664e2759eec2ba4ca64"} Dec 10 18:59:38 crc kubenswrapper[4894]: I1210 18:59:38.569817 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:38 crc kubenswrapper[4894]: I1210 18:59:38.569942 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:38 crc kubenswrapper[4894]: I1210 18:59:38.579358 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:38 crc kubenswrapper[4894]: I1210 18:59:38.877026 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" podUID="02ec8b17-2505-4d2b-a857-c92e5f47b484" containerName="oauth-openshift" containerID="cri-o://86f6aebe6299fbe188e18f87065f8123e3c5f839ec5c675a216585181059b5b4" gracePeriod=15 Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.229598 4894 generic.go:334] "Generic (PLEG): container finished" podID="02ec8b17-2505-4d2b-a857-c92e5f47b484" containerID="86f6aebe6299fbe188e18f87065f8123e3c5f839ec5c675a216585181059b5b4" exitCode=0 Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.229661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" event={"ID":"02ec8b17-2505-4d2b-a857-c92e5f47b484","Type":"ContainerDied","Data":"86f6aebe6299fbe188e18f87065f8123e3c5f839ec5c675a216585181059b5b4"} Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.485880 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524364 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524452 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-policies\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524555 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblkp\" (UniqueName: \"kubernetes.io/projected/02ec8b17-2505-4d2b-a857-c92e5f47b484-kube-api-access-nblkp\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524600 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-ocp-branding-template\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524641 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-dir\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524708 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-service-ca\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524747 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-idp-0-file-data\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-error\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524836 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.524957 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-router-certs\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.525030 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.525069 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-cliconfig\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.525115 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.525158 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") pod \"02ec8b17-2505-4d2b-a857-c92e5f47b484\" (UID: \"02ec8b17-2505-4d2b-a857-c92e5f47b484\") " Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.525715 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.525897 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.526260 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.526537 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.527204 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.532914 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.532963 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.533492 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.533676 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ec8b17-2505-4d2b-a857-c92e5f47b484-kube-api-access-nblkp" (OuterVolumeSpecName: "kube-api-access-nblkp") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "kube-api-access-nblkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.534020 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.534086 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.534125 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.534163 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.534541 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "02ec8b17-2505-4d2b-a857-c92e5f47b484" (UID: "02ec8b17-2505-4d2b-a857-c92e5f47b484"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.626940 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.626991 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627014 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627035 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627056 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblkp\" (UniqueName: \"kubernetes.io/projected/02ec8b17-2505-4d2b-a857-c92e5f47b484-kube-api-access-nblkp\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627098 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627119 4894 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02ec8b17-2505-4d2b-a857-c92e5f47b484-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627139 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627159 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627178 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627198 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627221 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627241 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:39 crc kubenswrapper[4894]: I1210 18:59:39.627260 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/02ec8b17-2505-4d2b-a857-c92e5f47b484-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:40 crc kubenswrapper[4894]: I1210 18:59:40.237674 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" event={"ID":"02ec8b17-2505-4d2b-a857-c92e5f47b484","Type":"ContainerDied","Data":"948289cdc42e6c91dd2a9cafe6322554abaad57551817f6b7c9067a8cf4e0db8"} Dec 10 18:59:40 crc kubenswrapper[4894]: I1210 18:59:40.237738 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m2s4d" Dec 10 18:59:40 crc kubenswrapper[4894]: I1210 18:59:40.238255 4894 scope.go:117] "RemoveContainer" containerID="86f6aebe6299fbe188e18f87065f8123e3c5f839ec5c675a216585181059b5b4" Dec 10 18:59:40 crc kubenswrapper[4894]: I1210 18:59:40.390533 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:41 crc kubenswrapper[4894]: I1210 18:59:41.217562 4894 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:41 crc kubenswrapper[4894]: I1210 18:59:41.342935 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f3647726-41b0-4977-8992-053db1835679" Dec 10 18:59:42 crc kubenswrapper[4894]: I1210 18:59:42.250692 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:42 crc kubenswrapper[4894]: I1210 18:59:42.251284 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:42 crc kubenswrapper[4894]: I1210 18:59:42.254894 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f3647726-41b0-4977-8992-053db1835679" Dec 10 18:59:42 crc kubenswrapper[4894]: I1210 18:59:42.263106 4894 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://a025f70968fae154e24c2fa7499838bf168c3991cd9366582c96c0594bc2b349" Dec 10 18:59:42 crc kubenswrapper[4894]: I1210 18:59:42.263146 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:43 crc kubenswrapper[4894]: I1210 18:59:43.257774 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:43 crc kubenswrapper[4894]: I1210 18:59:43.257824 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 18:59:43 crc kubenswrapper[4894]: I1210 18:59:43.261687 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f3647726-41b0-4977-8992-053db1835679" Dec 10 18:59:44 crc kubenswrapper[4894]: I1210 18:59:44.912926 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:44 crc kubenswrapper[4894]: I1210 18:59:44.920731 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:48 crc kubenswrapper[4894]: I1210 18:59:48.388840 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 18:59:48 crc kubenswrapper[4894]: I1210 18:59:48.478022 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:59:49 crc kubenswrapper[4894]: I1210 18:59:49.667100 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 18:59:50 crc kubenswrapper[4894]: I1210 18:59:50.404777 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:51 crc kubenswrapper[4894]: I1210 18:59:51.380754 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 18:59:51 crc kubenswrapper[4894]: I1210 18:59:51.930878 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 18:59:52 crc kubenswrapper[4894]: I1210 18:59:52.140127 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 18:59:52 crc kubenswrapper[4894]: I1210 18:59:52.319840 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 18:59:52 crc kubenswrapper[4894]: I1210 18:59:52.448911 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 18:59:52 crc kubenswrapper[4894]: I1210 18:59:52.865500 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4894]: I1210 18:59:53.023278 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 18:59:53 crc kubenswrapper[4894]: I1210 18:59:53.174505 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 18:59:53 crc kubenswrapper[4894]: I1210 18:59:53.600235 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 18:59:53 crc kubenswrapper[4894]: I1210 18:59:53.977755 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 18:59:53 crc kubenswrapper[4894]: I1210 18:59:53.998046 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 18:59:54 crc kubenswrapper[4894]: I1210 18:59:54.235882 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4894]: I1210 18:59:54.455531 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 18:59:54 crc kubenswrapper[4894]: I1210 18:59:54.657484 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4894]: I1210 18:59:54.799423 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4894]: I1210 18:59:54.975518 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.173602 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.367648 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.441567 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.645790 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.665556 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.685696 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 18:59:55 crc kubenswrapper[4894]: I1210 18:59:55.782652 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.028563 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.035224 4894 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.136728 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.165807 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.170266 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.234196 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.283190 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.352279 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.565484 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.621536 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.768606 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.858296 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.869780 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.928031 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.975375 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 18:59:56 crc kubenswrapper[4894]: I1210 18:59:56.986899 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.029010 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.042391 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.054046 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.067921 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.104475 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.301076 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.321501 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.406178 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.414885 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.631941 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.644747 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.663521 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.699731 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.803509 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.823493 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.855836 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4894]: I1210 18:59:57.979051 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.016414 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.049790 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.088636 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.100744 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.158936 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.174284 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.211579 4894 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.211861 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.216469 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.250090 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.353490 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.367362 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.370679 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.452475 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.473519 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.496988 4894 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.528895 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.580102 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.645315 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.667351 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.684957 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.764307 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.768600 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.781091 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.953111 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 18:59:58 crc kubenswrapper[4894]: I1210 18:59:58.979200 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.050964 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.087368 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.123630 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.179313 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.198327 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.252824 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.343384 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.466295 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.493810 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.535695 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.536452 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.610619 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.625415 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.652041 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.787488 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.857562 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.867438 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.891913 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.956765 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 18:59:59 crc kubenswrapper[4894]: I1210 18:59:59.979880 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.017467 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.019100 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.027348 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.037482 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.091390 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.115806 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.203385 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.290784 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.296428 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.317284 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.393750 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.402697 4894 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.409733 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.432189 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.440718 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.476169 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.506355 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.523311 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.606937 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.739429 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.835043 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 19:00:00 crc kubenswrapper[4894]: I1210 19:00:00.983669 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.059529 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.103052 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.196942 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.317814 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.418099 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.447504 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.511982 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.529479 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.536703 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.559490 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.608463 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.758781 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.790455 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.893048 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 19:00:01 crc kubenswrapper[4894]: I1210 19:00:01.931964 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.022321 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.036946 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.047319 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.203958 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.227266 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.267071 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.355302 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.395736 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.420136 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.863748 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.900915 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.941782 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 19:00:02 crc kubenswrapper[4894]: I1210 19:00:02.982715 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.001101 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.013961 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.526305 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.552014 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.575834 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.662942 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.734397 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.746203 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.933500 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.937362 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 19:00:03 crc kubenswrapper[4894]: I1210 19:00:03.992875 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.051266 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.095805 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.111547 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.153336 4894 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.157547 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.15752265 podStartE2EDuration="42.15752265s" podCreationTimestamp="2025-12-10 18:59:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:41.266865203 +0000 UTC m=+264.061713025" watchObservedRunningTime="2025-12-10 19:00:04.15752265 +0000 UTC m=+286.952370472" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.161318 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m2s4d","openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.161416 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-56c748df47-s5xc7","openshift-kube-apiserver/kube-apiserver-crc","openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp"] Dec 10 19:00:04 crc kubenswrapper[4894]: E1210 19:00:04.161752 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" containerName="installer" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.161786 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" containerName="installer" Dec 10 19:00:04 crc kubenswrapper[4894]: E1210 19:00:04.161814 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ec8b17-2505-4d2b-a857-c92e5f47b484" containerName="oauth-openshift" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.161816 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.161862 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="eb3fbf28-4903-4e44-8144-86c98da53f33" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.161832 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ec8b17-2505-4d2b-a857-c92e5f47b484" containerName="oauth-openshift" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.162205 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ec8b17-2505-4d2b-a857-c92e5f47b484" containerName="oauth-openshift" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.162245 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5bc823d-aae9-4d4e-b6cd-095590924c6c" containerName="installer" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.163104 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.165563 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.167225 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.167578 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.168086 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.168519 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.168696 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.167636 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.169571 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.169772 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.170202 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.171560 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.174324 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.174439 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.174325 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.175106 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.176515 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.181181 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.183790 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.190939 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.209072 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.214695 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.217120 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.219798 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.219770166 podStartE2EDuration="23.219770166s" podCreationTimestamp="2025-12-10 18:59:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:00:04.200171556 +0000 UTC m=+286.995019418" watchObservedRunningTime="2025-12-10 19:00:04.219770166 +0000 UTC m=+287.014617968" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.233438 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250489 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vh5q\" (UniqueName: \"kubernetes.io/projected/6e789ccb-359b-4c16-a8f5-25864f34778c-kube-api-access-8vh5q\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250531 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250556 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-audit-policies\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250574 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250590 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e789ccb-359b-4c16-a8f5-25864f34778c-config-volume\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250607 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt989\" (UniqueName: \"kubernetes.io/projected/e073fba1-9a04-4a93-8983-ebcec45567a9-kube-api-access-rt989\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250653 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-login\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250670 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-router-certs\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250685 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e789ccb-359b-4c16-a8f5-25864f34778c-secret-volume\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250700 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e073fba1-9a04-4a93-8983-ebcec45567a9-audit-dir\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250716 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250741 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250756 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-session\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250773 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-error\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250794 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-service-ca\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.250812 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.274739 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.310063 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.346085 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351808 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-service-ca\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351864 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vh5q\" (UniqueName: \"kubernetes.io/projected/6e789ccb-359b-4c16-a8f5-25864f34778c-kube-api-access-8vh5q\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351918 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-audit-policies\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351961 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351978 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e789ccb-359b-4c16-a8f5-25864f34778c-config-volume\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.351996 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352020 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt989\" (UniqueName: \"kubernetes.io/projected/e073fba1-9a04-4a93-8983-ebcec45567a9-kube-api-access-rt989\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352045 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-login\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352061 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-router-certs\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352229 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e789ccb-359b-4c16-a8f5-25864f34778c-secret-volume\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352244 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e073fba1-9a04-4a93-8983-ebcec45567a9-audit-dir\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352260 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352287 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-session\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-error\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.352995 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.353197 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e789ccb-359b-4c16-a8f5-25864f34778c-config-volume\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.353377 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-service-ca\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.353565 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-audit-policies\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.353874 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e073fba1-9a04-4a93-8983-ebcec45567a9-audit-dir\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.353901 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.357507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.357956 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-session\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.358431 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-error\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.360893 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.361741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.361868 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.362401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-system-router-certs\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.363410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e789ccb-359b-4c16-a8f5-25864f34778c-secret-volume\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.364148 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e073fba1-9a04-4a93-8983-ebcec45567a9-v4-0-config-user-template-login\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.369270 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt989\" (UniqueName: \"kubernetes.io/projected/e073fba1-9a04-4a93-8983-ebcec45567a9-kube-api-access-rt989\") pod \"oauth-openshift-56c748df47-s5xc7\" (UID: \"e073fba1-9a04-4a93-8983-ebcec45567a9\") " pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.378761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vh5q\" (UniqueName: \"kubernetes.io/projected/6e789ccb-359b-4c16-a8f5-25864f34778c-kube-api-access-8vh5q\") pod \"collect-profiles-29423220-2zvnp\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.437639 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.503656 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.520147 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.602757 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.634259 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.712643 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.720968 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.721202 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.780211 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.792229 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.806528 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.808433 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.842653 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.888728 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.901749 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 19:00:04 crc kubenswrapper[4894]: I1210 19:00:04.967269 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.008665 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.047269 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.081057 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.171954 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.550959 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ec8b17-2505-4d2b-a857-c92e5f47b484" path="/var/lib/kubelet/pods/02ec8b17-2505-4d2b-a857-c92e5f47b484/volumes" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.582198 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.634163 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.748365 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.758021 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.832217 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.873360 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.911171 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 19:00:05 crc kubenswrapper[4894]: I1210 19:00:05.931462 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.081722 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.117947 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.122195 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.176967 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.246021 4894 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.502881 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.511410 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.533127 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.571432 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.628548 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp"] Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.634043 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56c748df47-s5xc7"] Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.775912 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.793347 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.909897 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56c748df47-s5xc7"] Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.918624 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.937660 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.985830 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 19:00:06 crc kubenswrapper[4894]: I1210 19:00:06.994987 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp"] Dec 10 19:00:07 crc kubenswrapper[4894]: W1210 19:00:07.002092 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e789ccb_359b_4c16_a8f5_25864f34778c.slice/crio-d2114c5992d917107c947fb23c6897ae0a10a7eb91bf307cb3235afbce01b900 WatchSource:0}: Error finding container d2114c5992d917107c947fb23c6897ae0a10a7eb91bf307cb3235afbce01b900: Status 404 returned error can't find the container with id d2114c5992d917107c947fb23c6897ae0a10a7eb91bf307cb3235afbce01b900 Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.052058 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.076980 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.081734 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.089452 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.185882 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.199764 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.404983 4894 generic.go:334] "Generic (PLEG): container finished" podID="6e789ccb-359b-4c16-a8f5-25864f34778c" containerID="efd9770e718f824bf8d557c632532fb5fadb807e233864fca99663a3563c4df9" exitCode=0 Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.405086 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" event={"ID":"6e789ccb-359b-4c16-a8f5-25864f34778c","Type":"ContainerDied","Data":"efd9770e718f824bf8d557c632532fb5fadb807e233864fca99663a3563c4df9"} Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.405362 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" event={"ID":"6e789ccb-359b-4c16-a8f5-25864f34778c","Type":"ContainerStarted","Data":"d2114c5992d917107c947fb23c6897ae0a10a7eb91bf307cb3235afbce01b900"} Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.406596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" event={"ID":"e073fba1-9a04-4a93-8983-ebcec45567a9","Type":"ContainerStarted","Data":"79c296d3aac6bae49913ad111491b2e34c6880a5368f7c7e45f345d099dd9d5b"} Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.406636 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" event={"ID":"e073fba1-9a04-4a93-8983-ebcec45567a9","Type":"ContainerStarted","Data":"8ff440febba39f2742eaa1993206542b5c59da282c211d716ed50e7373adf09c"} Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.406870 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.440540 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" podStartSLOduration=54.440518029 podStartE2EDuration="54.440518029s" podCreationTimestamp="2025-12-10 18:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:00:07.438013398 +0000 UTC m=+290.232861210" watchObservedRunningTime="2025-12-10 19:00:07.440518029 +0000 UTC m=+290.235365821" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.458222 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-56c748df47-s5xc7" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.546886 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.676710 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 19:00:07 crc kubenswrapper[4894]: I1210 19:00:07.776448 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.036115 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.109083 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.169984 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.224305 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.326455 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.491571 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.706261 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.813410 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e789ccb-359b-4c16-a8f5-25864f34778c-secret-volume\") pod \"6e789ccb-359b-4c16-a8f5-25864f34778c\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.813546 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vh5q\" (UniqueName: \"kubernetes.io/projected/6e789ccb-359b-4c16-a8f5-25864f34778c-kube-api-access-8vh5q\") pod \"6e789ccb-359b-4c16-a8f5-25864f34778c\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.813595 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e789ccb-359b-4c16-a8f5-25864f34778c-config-volume\") pod \"6e789ccb-359b-4c16-a8f5-25864f34778c\" (UID: \"6e789ccb-359b-4c16-a8f5-25864f34778c\") " Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.814344 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e789ccb-359b-4c16-a8f5-25864f34778c-config-volume" (OuterVolumeSpecName: "config-volume") pod "6e789ccb-359b-4c16-a8f5-25864f34778c" (UID: "6e789ccb-359b-4c16-a8f5-25864f34778c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.820322 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e789ccb-359b-4c16-a8f5-25864f34778c-kube-api-access-8vh5q" (OuterVolumeSpecName: "kube-api-access-8vh5q") pod "6e789ccb-359b-4c16-a8f5-25864f34778c" (UID: "6e789ccb-359b-4c16-a8f5-25864f34778c"). InnerVolumeSpecName "kube-api-access-8vh5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.821529 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e789ccb-359b-4c16-a8f5-25864f34778c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6e789ccb-359b-4c16-a8f5-25864f34778c" (UID: "6e789ccb-359b-4c16-a8f5-25864f34778c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.848159 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.902398 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.914454 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e789ccb-359b-4c16-a8f5-25864f34778c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.914496 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e789ccb-359b-4c16-a8f5-25864f34778c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.914514 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vh5q\" (UniqueName: \"kubernetes.io/projected/6e789ccb-359b-4c16-a8f5-25864f34778c-kube-api-access-8vh5q\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:08 crc kubenswrapper[4894]: I1210 19:00:08.996632 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 19:00:09 crc kubenswrapper[4894]: I1210 19:00:09.108299 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 19:00:09 crc kubenswrapper[4894]: I1210 19:00:09.419686 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" event={"ID":"6e789ccb-359b-4c16-a8f5-25864f34778c","Type":"ContainerDied","Data":"d2114c5992d917107c947fb23c6897ae0a10a7eb91bf307cb3235afbce01b900"} Dec 10 19:00:09 crc kubenswrapper[4894]: I1210 19:00:09.419732 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2114c5992d917107c947fb23c6897ae0a10a7eb91bf307cb3235afbce01b900" Dec 10 19:00:09 crc kubenswrapper[4894]: I1210 19:00:09.419756 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp" Dec 10 19:00:11 crc kubenswrapper[4894]: I1210 19:00:11.083520 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 19:00:15 crc kubenswrapper[4894]: I1210 19:00:15.185566 4894 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 19:00:15 crc kubenswrapper[4894]: I1210 19:00:15.186398 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://21e93b9b72eb748ce7e5d3ec955009922ec5e52c59d77dd90392a737369e18b2" gracePeriod=5 Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.491917 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.492473 4894 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="21e93b9b72eb748ce7e5d3ec955009922ec5e52c59d77dd90392a737369e18b2" exitCode=137 Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.823370 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.823471 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891339 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891378 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891435 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891544 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891583 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891608 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891600 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891658 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891828 4894 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891839 4894 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891937 4894 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.891953 4894 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.899892 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:20 crc kubenswrapper[4894]: I1210 19:00:20.992789 4894 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.499132 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.499294 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.499331 4894 scope.go:117] "RemoveContainer" containerID="21e93b9b72eb748ce7e5d3ec955009922ec5e52c59d77dd90392a737369e18b2" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.547670 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.547902 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.566885 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.566916 4894 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="113286f3-d80e-4783-aca7-df4cbd72ab3f" Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.571227 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 19:00:21 crc kubenswrapper[4894]: I1210 19:00:21.571244 4894 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="113286f3-d80e-4783-aca7-df4cbd72ab3f" Dec 10 19:00:30 crc kubenswrapper[4894]: I1210 19:00:30.557768 4894 generic.go:334] "Generic (PLEG): container finished" podID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerID="838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c" exitCode=0 Dec 10 19:00:30 crc kubenswrapper[4894]: I1210 19:00:30.557914 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" event={"ID":"600e8028-f30b-423a-b0fc-c5f461d5fea3","Type":"ContainerDied","Data":"838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c"} Dec 10 19:00:30 crc kubenswrapper[4894]: I1210 19:00:30.559400 4894 scope.go:117] "RemoveContainer" containerID="838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c" Dec 10 19:00:31 crc kubenswrapper[4894]: I1210 19:00:31.568988 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" event={"ID":"600e8028-f30b-423a-b0fc-c5f461d5fea3","Type":"ContainerStarted","Data":"9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048"} Dec 10 19:00:31 crc kubenswrapper[4894]: I1210 19:00:31.570018 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 19:00:31 crc kubenswrapper[4894]: I1210 19:00:31.575239 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.280081 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kh2f7"] Dec 10 19:00:57 crc kubenswrapper[4894]: E1210 19:00:57.281195 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.281231 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 19:00:57 crc kubenswrapper[4894]: E1210 19:00:57.281283 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e789ccb-359b-4c16-a8f5-25864f34778c" containerName="collect-profiles" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.281296 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e789ccb-359b-4c16-a8f5-25864f34778c" containerName="collect-profiles" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.281500 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e789ccb-359b-4c16-a8f5-25864f34778c" containerName="collect-profiles" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.281512 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.282067 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.294612 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kh2f7"] Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.406608 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-trusted-ca\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.406656 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-registry-certificates\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.406703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.406719 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.406844 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmk9j\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-kube-api-access-jmk9j\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.406920 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.407022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-bound-sa-token\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.407061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-registry-tls\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.433788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.507704 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508159 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmk9j\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-kube-api-access-jmk9j\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-bound-sa-token\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508214 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-registry-tls\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508239 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-trusted-ca\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508261 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-registry-certificates\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.508329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.509434 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-registry-certificates\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.510185 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-trusted-ca\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.513593 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-registry-tls\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.513602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.525738 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmk9j\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-kube-api-access-jmk9j\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.527608 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce-bound-sa-token\") pod \"image-registry-66df7c8f76-kh2f7\" (UID: \"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce\") " pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:57 crc kubenswrapper[4894]: I1210 19:00:57.616191 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:58 crc kubenswrapper[4894]: I1210 19:00:58.037925 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kh2f7"] Dec 10 19:00:58 crc kubenswrapper[4894]: W1210 19:00:58.049764 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedd5fa24_8107_4d2b_b9d2_6bb4f90e5dce.slice/crio-6d0cbd35a0cdfa47855e35f5ed4a338c14cfc6a5f0b3bd7a4e85fd57db281e57 WatchSource:0}: Error finding container 6d0cbd35a0cdfa47855e35f5ed4a338c14cfc6a5f0b3bd7a4e85fd57db281e57: Status 404 returned error can't find the container with id 6d0cbd35a0cdfa47855e35f5ed4a338c14cfc6a5f0b3bd7a4e85fd57db281e57 Dec 10 19:00:58 crc kubenswrapper[4894]: I1210 19:00:58.755556 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" event={"ID":"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce","Type":"ContainerStarted","Data":"81b9d8d7f6a8c5ea5df49f8559236f39d19a7e420d9f22af925a242ccb1a3215"} Dec 10 19:00:58 crc kubenswrapper[4894]: I1210 19:00:58.756179 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:00:58 crc kubenswrapper[4894]: I1210 19:00:58.756213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" event={"ID":"edd5fa24-8107-4d2b-b9d2-6bb4f90e5dce","Type":"ContainerStarted","Data":"6d0cbd35a0cdfa47855e35f5ed4a338c14cfc6a5f0b3bd7a4e85fd57db281e57"} Dec 10 19:00:58 crc kubenswrapper[4894]: I1210 19:00:58.779240 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" podStartSLOduration=1.779205589 podStartE2EDuration="1.779205589s" podCreationTimestamp="2025-12-10 19:00:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:00:58.775075499 +0000 UTC m=+341.569923381" watchObservedRunningTime="2025-12-10 19:00:58.779205589 +0000 UTC m=+341.574053431" Dec 10 19:01:17 crc kubenswrapper[4894]: I1210 19:01:17.620236 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-kh2f7" Dec 10 19:01:17 crc kubenswrapper[4894]: I1210 19:01:17.681885 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jnckv"] Dec 10 19:01:19 crc kubenswrapper[4894]: I1210 19:01:19.577953 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:01:19 crc kubenswrapper[4894]: I1210 19:01:19.578522 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.454257 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zdvpp"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.455164 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zdvpp" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="registry-server" containerID="cri-o://13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492" gracePeriod=30 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.462838 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vr2zg"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.464231 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vr2zg" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="registry-server" containerID="cri-o://0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b" gracePeriod=30 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.485760 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jxj6j"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.486031 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" containerID="cri-o://9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048" gracePeriod=30 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.490103 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mb95"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.490343 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9mb95" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="registry-server" containerID="cri-o://6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc" gracePeriod=30 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.508544 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vnn24"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.508966 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vnn24" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="registry-server" containerID="cri-o://fa3bd2137101ef56eed2e6b51530d50e3b3b39ba5335c6e7eb7f923dc0bb991c" gracePeriod=30 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.512719 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sgnhf"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.513893 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.521531 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.521580 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.521603 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmrkc\" (UniqueName: \"kubernetes.io/projected/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-kube-api-access-cmrkc\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.525723 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sgnhf"] Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.593474 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jxj6j container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.593527 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.622588 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.622655 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.622690 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmrkc\" (UniqueName: \"kubernetes.io/projected/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-kube-api-access-cmrkc\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.624909 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.637815 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.645604 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmrkc\" (UniqueName: \"kubernetes.io/projected/8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea-kube-api-access-cmrkc\") pod \"marketplace-operator-79b997595-sgnhf\" (UID: \"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea\") " pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.872711 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.877735 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.914921 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.921348 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.922773 4894 generic.go:334] "Generic (PLEG): container finished" podID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerID="13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492" exitCode=0 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.922808 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.922836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdvpp" event={"ID":"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee","Type":"ContainerDied","Data":"13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.922875 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdvpp" event={"ID":"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee","Type":"ContainerDied","Data":"3b2857ba49a15d2f148e5669d25e0da8f7bab92660e4d7b6928825a0a948d6d0"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.922892 4894 scope.go:117] "RemoveContainer" containerID="13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.922948 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdvpp" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.926260 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca\") pod \"600e8028-f30b-423a-b0fc-c5f461d5fea3\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.927294 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "600e8028-f30b-423a-b0fc-c5f461d5fea3" (UID: "600e8028-f30b-423a-b0fc-c5f461d5fea3"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.927792 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-catalog-content\") pod \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.927831 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-utilities\") pod \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.927886 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfs9l\" (UniqueName: \"kubernetes.io/projected/600e8028-f30b-423a-b0fc-c5f461d5fea3-kube-api-access-tfs9l\") pod \"600e8028-f30b-423a-b0fc-c5f461d5fea3\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.927920 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-catalog-content\") pod \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.927978 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5867z\" (UniqueName: \"kubernetes.io/projected/91c2b024-f9b4-4d85-bc5d-ad463ae51871-kube-api-access-5867z\") pod \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.928006 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-utilities\") pod \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.928029 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb52m\" (UniqueName: \"kubernetes.io/projected/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-kube-api-access-cb52m\") pod \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\" (UID: \"2a2f64bf-a1f8-4ad3-9dd3-970afc101879\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.928058 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics\") pod \"600e8028-f30b-423a-b0fc-c5f461d5fea3\" (UID: \"600e8028-f30b-423a-b0fc-c5f461d5fea3\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.928087 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzbw7\" (UniqueName: \"kubernetes.io/projected/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-kube-api-access-mzbw7\") pod \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.928113 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-utilities\") pod \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\" (UID: \"91c2b024-f9b4-4d85-bc5d-ad463ae51871\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.928139 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-catalog-content\") pod \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\" (UID: \"b845e384-badb-4fbe-92c5-9ba9b9b8a8ee\") " Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.929722 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.931330 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-utilities" (OuterVolumeSpecName: "utilities") pod "2a2f64bf-a1f8-4ad3-9dd3-970afc101879" (UID: "2a2f64bf-a1f8-4ad3-9dd3-970afc101879"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.935048 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-utilities" (OuterVolumeSpecName: "utilities") pod "b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" (UID: "b845e384-badb-4fbe-92c5-9ba9b9b8a8ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.935400 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerID="0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b" exitCode=0 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.935487 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2zg" event={"ID":"2a2f64bf-a1f8-4ad3-9dd3-970afc101879","Type":"ContainerDied","Data":"0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.935519 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2zg" event={"ID":"2a2f64bf-a1f8-4ad3-9dd3-970afc101879","Type":"ContainerDied","Data":"24376163083bc8fd9c2583bf90caf3f99078d21c2403c9fb9d29d7fee5a8da4c"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.938717 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-kube-api-access-cb52m" (OuterVolumeSpecName: "kube-api-access-cb52m") pod "2a2f64bf-a1f8-4ad3-9dd3-970afc101879" (UID: "2a2f64bf-a1f8-4ad3-9dd3-970afc101879"). InnerVolumeSpecName "kube-api-access-cb52m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.938866 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2zg" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.939450 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-utilities" (OuterVolumeSpecName: "utilities") pod "91c2b024-f9b4-4d85-bc5d-ad463ae51871" (UID: "91c2b024-f9b4-4d85-bc5d-ad463ae51871"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.943224 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-kube-api-access-mzbw7" (OuterVolumeSpecName: "kube-api-access-mzbw7") pod "b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" (UID: "b845e384-badb-4fbe-92c5-9ba9b9b8a8ee"). InnerVolumeSpecName "kube-api-access-mzbw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.946804 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91c2b024-f9b4-4d85-bc5d-ad463ae51871-kube-api-access-5867z" (OuterVolumeSpecName: "kube-api-access-5867z") pod "91c2b024-f9b4-4d85-bc5d-ad463ae51871" (UID: "91c2b024-f9b4-4d85-bc5d-ad463ae51871"). InnerVolumeSpecName "kube-api-access-5867z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.947224 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/600e8028-f30b-423a-b0fc-c5f461d5fea3-kube-api-access-tfs9l" (OuterVolumeSpecName: "kube-api-access-tfs9l") pod "600e8028-f30b-423a-b0fc-c5f461d5fea3" (UID: "600e8028-f30b-423a-b0fc-c5f461d5fea3"). InnerVolumeSpecName "kube-api-access-tfs9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.952747 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "600e8028-f30b-423a-b0fc-c5f461d5fea3" (UID: "600e8028-f30b-423a-b0fc-c5f461d5fea3"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.955180 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerDied","Data":"fa3bd2137101ef56eed2e6b51530d50e3b3b39ba5335c6e7eb7f923dc0bb991c"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.955012 4894 generic.go:334] "Generic (PLEG): container finished" podID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerID="fa3bd2137101ef56eed2e6b51530d50e3b3b39ba5335c6e7eb7f923dc0bb991c" exitCode=0 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.965663 4894 scope.go:117] "RemoveContainer" containerID="62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.974818 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91c2b024-f9b4-4d85-bc5d-ad463ae51871" (UID: "91c2b024-f9b4-4d85-bc5d-ad463ae51871"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.991517 4894 generic.go:334] "Generic (PLEG): container finished" podID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerID="6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc" exitCode=0 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.991601 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mb95" event={"ID":"91c2b024-f9b4-4d85-bc5d-ad463ae51871","Type":"ContainerDied","Data":"6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.991627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mb95" event={"ID":"91c2b024-f9b4-4d85-bc5d-ad463ae51871","Type":"ContainerDied","Data":"e2ec6c230fa29dd389472d457c92c6a3e7e02da42c3c1225a7fbcb83445b468d"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.991679 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mb95" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.994036 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.997556 4894 scope.go:117] "RemoveContainer" containerID="5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738" Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.999135 4894 generic.go:334] "Generic (PLEG): container finished" podID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerID="9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048" exitCode=0 Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.999174 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" event={"ID":"600e8028-f30b-423a-b0fc-c5f461d5fea3","Type":"ContainerDied","Data":"9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.999196 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" event={"ID":"600e8028-f30b-423a-b0fc-c5f461d5fea3","Type":"ContainerDied","Data":"17b8dad71997b60bc4bf0f33ee4b9a673c590880d1669749e093a7b7745f590c"} Dec 10 19:01:24 crc kubenswrapper[4894]: I1210 19:01:24.999233 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jxj6j" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.004234 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a2f64bf-a1f8-4ad3-9dd3-970afc101879" (UID: "2a2f64bf-a1f8-4ad3-9dd3-970afc101879"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.015640 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" (UID: "b845e384-badb-4fbe-92c5-9ba9b9b8a8ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.017755 4894 scope.go:117] "RemoveContainer" containerID="13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.018694 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492\": container with ID starting with 13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492 not found: ID does not exist" containerID="13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.018724 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492"} err="failed to get container status \"13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492\": rpc error: code = NotFound desc = could not find container \"13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492\": container with ID starting with 13b3e4c9383a9ef2d8e47a33f73daf3a6621db57799aa8b73708f9726ac94492 not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.018749 4894 scope.go:117] "RemoveContainer" containerID="62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.019143 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a\": container with ID starting with 62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a not found: ID does not exist" containerID="62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.019184 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a"} err="failed to get container status \"62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a\": rpc error: code = NotFound desc = could not find container \"62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a\": container with ID starting with 62c47b1c582f278112d0107c196eaa2ae3babf650d1fdabb5e64858b2a457c9a not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.019220 4894 scope.go:117] "RemoveContainer" containerID="5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.023661 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738\": container with ID starting with 5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738 not found: ID does not exist" containerID="5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.023700 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738"} err="failed to get container status \"5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738\": rpc error: code = NotFound desc = could not find container \"5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738\": container with ID starting with 5c3c80fe3c3b9d11c0d614d06afe98c3f3dc5e8da6ac333dc8b108e571741738 not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.023723 4894 scope.go:117] "RemoveContainer" containerID="0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032053 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032154 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032168 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032177 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032187 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfs9l\" (UniqueName: \"kubernetes.io/projected/600e8028-f30b-423a-b0fc-c5f461d5fea3-kube-api-access-tfs9l\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032196 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91c2b024-f9b4-4d85-bc5d-ad463ae51871-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032220 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5867z\" (UniqueName: \"kubernetes.io/projected/91c2b024-f9b4-4d85-bc5d-ad463ae51871-kube-api-access-5867z\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032231 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032239 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb52m\" (UniqueName: \"kubernetes.io/projected/2a2f64bf-a1f8-4ad3-9dd3-970afc101879-kube-api-access-cb52m\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032249 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/600e8028-f30b-423a-b0fc-c5f461d5fea3-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.032259 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzbw7\" (UniqueName: \"kubernetes.io/projected/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee-kube-api-access-mzbw7\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.049956 4894 scope.go:117] "RemoveContainer" containerID="95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.075292 4894 scope.go:117] "RemoveContainer" containerID="0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.078917 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jxj6j"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.081804 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jxj6j"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.086233 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mb95"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.087561 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mb95"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.093349 4894 scope.go:117] "RemoveContainer" containerID="0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.093815 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b\": container with ID starting with 0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b not found: ID does not exist" containerID="0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.093855 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b"} err="failed to get container status \"0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b\": rpc error: code = NotFound desc = could not find container \"0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b\": container with ID starting with 0035a7a6f938b4d4897f4777543a7f3c93f7d0123194db939af9e766e4f8f01b not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.093877 4894 scope.go:117] "RemoveContainer" containerID="95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.097177 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7\": container with ID starting with 95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7 not found: ID does not exist" containerID="95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.097200 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7"} err="failed to get container status \"95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7\": rpc error: code = NotFound desc = could not find container \"95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7\": container with ID starting with 95cf99e8368f81ad4d51926311ec43ebc5760a06275a2b445341a88926f4aca7 not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.097216 4894 scope.go:117] "RemoveContainer" containerID="0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.097637 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5\": container with ID starting with 0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5 not found: ID does not exist" containerID="0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.097661 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5"} err="failed to get container status \"0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5\": rpc error: code = NotFound desc = could not find container \"0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5\": container with ID starting with 0f752af3abbe439a95d096f118267c3cb6568319e329945e54cf1f90de44d5e5 not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.097676 4894 scope.go:117] "RemoveContainer" containerID="6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.111932 4894 scope.go:117] "RemoveContainer" containerID="d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.132335 4894 scope.go:117] "RemoveContainer" containerID="81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.132803 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-catalog-content\") pod \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.132939 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-utilities\") pod \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.132987 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk8l4\" (UniqueName: \"kubernetes.io/projected/a5e7b198-78b1-41d4-86e9-583e68ae92f1-kube-api-access-pk8l4\") pod \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\" (UID: \"a5e7b198-78b1-41d4-86e9-583e68ae92f1\") " Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.134362 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-utilities" (OuterVolumeSpecName: "utilities") pod "a5e7b198-78b1-41d4-86e9-583e68ae92f1" (UID: "a5e7b198-78b1-41d4-86e9-583e68ae92f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.138352 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e7b198-78b1-41d4-86e9-583e68ae92f1-kube-api-access-pk8l4" (OuterVolumeSpecName: "kube-api-access-pk8l4") pod "a5e7b198-78b1-41d4-86e9-583e68ae92f1" (UID: "a5e7b198-78b1-41d4-86e9-583e68ae92f1"). InnerVolumeSpecName "kube-api-access-pk8l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.148392 4894 scope.go:117] "RemoveContainer" containerID="6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.149587 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc\": container with ID starting with 6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc not found: ID does not exist" containerID="6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.149621 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc"} err="failed to get container status \"6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc\": rpc error: code = NotFound desc = could not find container \"6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc\": container with ID starting with 6ee31cbb0ed228f777557665cf150da7b010ee0087799dac9225b4d93ec1defc not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.149646 4894 scope.go:117] "RemoveContainer" containerID="d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.150157 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb\": container with ID starting with d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb not found: ID does not exist" containerID="d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.150234 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb"} err="failed to get container status \"d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb\": rpc error: code = NotFound desc = could not find container \"d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb\": container with ID starting with d2e613adb88efe4b478877c3f4144d3786ec3cb293c04f6240887e27e7266fcb not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.150270 4894 scope.go:117] "RemoveContainer" containerID="81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.150651 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a\": container with ID starting with 81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a not found: ID does not exist" containerID="81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.150687 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a"} err="failed to get container status \"81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a\": rpc error: code = NotFound desc = could not find container \"81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a\": container with ID starting with 81a61fe877c3aa9b9385b11a7ced0c87fa2ac7b2a71b4ea3a5291a671ab0fe9a not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.150707 4894 scope.go:117] "RemoveContainer" containerID="9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.165570 4894 scope.go:117] "RemoveContainer" containerID="838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.178046 4894 scope.go:117] "RemoveContainer" containerID="9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.178330 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048\": container with ID starting with 9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048 not found: ID does not exist" containerID="9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.178358 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048"} err="failed to get container status \"9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048\": rpc error: code = NotFound desc = could not find container \"9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048\": container with ID starting with 9f0737b9b2656ef0944ba2ea0034e53b9aa8230a91d9329860831d5a687c3048 not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.178382 4894 scope.go:117] "RemoveContainer" containerID="838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c" Dec 10 19:01:25 crc kubenswrapper[4894]: E1210 19:01:25.178625 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c\": container with ID starting with 838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c not found: ID does not exist" containerID="838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.178649 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c"} err="failed to get container status \"838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c\": rpc error: code = NotFound desc = could not find container \"838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c\": container with ID starting with 838b3cabe00809adf00971dccb8f0f26be454059bd27a89e43041072e710e37c not found: ID does not exist" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.234454 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk8l4\" (UniqueName: \"kubernetes.io/projected/a5e7b198-78b1-41d4-86e9-583e68ae92f1-kube-api-access-pk8l4\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.234483 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.249365 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5e7b198-78b1-41d4-86e9-583e68ae92f1" (UID: "a5e7b198-78b1-41d4-86e9-583e68ae92f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.255710 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zdvpp"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.261586 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zdvpp"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.283442 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vr2zg"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.290978 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vr2zg"] Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.335281 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e7b198-78b1-41d4-86e9-583e68ae92f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.406029 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sgnhf"] Dec 10 19:01:25 crc kubenswrapper[4894]: W1210 19:01:25.416029 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c9b37fd_63d5_4b44_9e8e_d41b6ec233ea.slice/crio-ad2315020f588c231c0a97f20b6766fc23e1bdec003fe67d0b0b154e80469f3d WatchSource:0}: Error finding container ad2315020f588c231c0a97f20b6766fc23e1bdec003fe67d0b0b154e80469f3d: Status 404 returned error can't find the container with id ad2315020f588c231c0a97f20b6766fc23e1bdec003fe67d0b0b154e80469f3d Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.552339 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" path="/var/lib/kubelet/pods/2a2f64bf-a1f8-4ad3-9dd3-970afc101879/volumes" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.554069 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" path="/var/lib/kubelet/pods/600e8028-f30b-423a-b0fc-c5f461d5fea3/volumes" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.555136 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" path="/var/lib/kubelet/pods/91c2b024-f9b4-4d85-bc5d-ad463ae51871/volumes" Dec 10 19:01:25 crc kubenswrapper[4894]: I1210 19:01:25.556388 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" path="/var/lib/kubelet/pods/b845e384-badb-4fbe-92c5-9ba9b9b8a8ee/volumes" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.005669 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" event={"ID":"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea","Type":"ContainerStarted","Data":"ee98f13c30ba19727bbdea1ded38300aa15023e6450d93a23fd9af50b1fa7f25"} Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.005710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" event={"ID":"8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea","Type":"ContainerStarted","Data":"ad2315020f588c231c0a97f20b6766fc23e1bdec003fe67d0b0b154e80469f3d"} Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.005998 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.008166 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnn24" event={"ID":"a5e7b198-78b1-41d4-86e9-583e68ae92f1","Type":"ContainerDied","Data":"ac03c2aaebbe519b9d6bc0a8b73b7c8e84c014eb63e4d2c773c72d3a58881fc4"} Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.008214 4894 scope.go:117] "RemoveContainer" containerID="fa3bd2137101ef56eed2e6b51530d50e3b3b39ba5335c6e7eb7f923dc0bb991c" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.008172 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnn24" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.012280 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.027653 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sgnhf" podStartSLOduration=2.027632588 podStartE2EDuration="2.027632588s" podCreationTimestamp="2025-12-10 19:01:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:01:26.025143012 +0000 UTC m=+368.819990804" watchObservedRunningTime="2025-12-10 19:01:26.027632588 +0000 UTC m=+368.822480380" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.028906 4894 scope.go:117] "RemoveContainer" containerID="2631b3936979dfc11e34d3fa0a5b685ad3958937e258efe91ac8468c62201d91" Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.035512 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vnn24"] Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.039999 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vnn24"] Dec 10 19:01:26 crc kubenswrapper[4894]: I1210 19:01:26.056632 4894 scope.go:117] "RemoveContainer" containerID="364da9cc0eb2f9e763d018fccbf352e53f2eafecf5ec65d99a9b101acdc919d5" Dec 10 19:01:27 crc kubenswrapper[4894]: I1210 19:01:27.551180 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" path="/var/lib/kubelet/pods/a5e7b198-78b1-41d4-86e9-583e68ae92f1/volumes" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.356391 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4lrz"] Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357294 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357308 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357318 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357326 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357337 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357344 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357354 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357361 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357370 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357377 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357387 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357394 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357402 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357409 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357426 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357433 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357443 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357450 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357458 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357465 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357477 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357484 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357491 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357498 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="extract-utilities" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357507 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357664 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="extract-content" Dec 10 19:01:36 crc kubenswrapper[4894]: E1210 19:01:36.357673 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357680 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357785 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="91c2b024-f9b4-4d85-bc5d-ad463ae51871" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357799 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a2f64bf-a1f8-4ad3-9dd3-970afc101879" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357809 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b845e384-badb-4fbe-92c5-9ba9b9b8a8ee" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357823 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357834 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="600e8028-f30b-423a-b0fc-c5f461d5fea3" containerName="marketplace-operator" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.357861 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e7b198-78b1-41d4-86e9-583e68ae92f1" containerName="registry-server" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.358696 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.361331 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.372641 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4lrz"] Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.383978 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6306ae09-754d-4285-8182-4f6433405855-catalog-content\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.384067 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6306ae09-754d-4285-8182-4f6433405855-utilities\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.384101 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/6306ae09-754d-4285-8182-4f6433405855-kube-api-access-2wrwl\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.484857 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6306ae09-754d-4285-8182-4f6433405855-utilities\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.484917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/6306ae09-754d-4285-8182-4f6433405855-kube-api-access-2wrwl\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.484966 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6306ae09-754d-4285-8182-4f6433405855-catalog-content\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.485468 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6306ae09-754d-4285-8182-4f6433405855-utilities\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.485478 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6306ae09-754d-4285-8182-4f6433405855-catalog-content\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.511296 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/6306ae09-754d-4285-8182-4f6433405855-kube-api-access-2wrwl\") pod \"certified-operators-v4lrz\" (UID: \"6306ae09-754d-4285-8182-4f6433405855\") " pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.544970 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c944h"] Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.549594 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.555060 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.570635 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c944h"] Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.585934 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42b609b-dba3-484f-b919-a33fc18e3385-catalog-content\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.586036 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x6s8\" (UniqueName: \"kubernetes.io/projected/d42b609b-dba3-484f-b919-a33fc18e3385-kube-api-access-7x6s8\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.586222 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42b609b-dba3-484f-b919-a33fc18e3385-utilities\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.676271 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.687558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42b609b-dba3-484f-b919-a33fc18e3385-utilities\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.688093 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42b609b-dba3-484f-b919-a33fc18e3385-catalog-content\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.688203 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x6s8\" (UniqueName: \"kubernetes.io/projected/d42b609b-dba3-484f-b919-a33fc18e3385-kube-api-access-7x6s8\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.688361 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42b609b-dba3-484f-b919-a33fc18e3385-utilities\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.688647 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42b609b-dba3-484f-b919-a33fc18e3385-catalog-content\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.707279 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x6s8\" (UniqueName: \"kubernetes.io/projected/d42b609b-dba3-484f-b919-a33fc18e3385-kube-api-access-7x6s8\") pod \"community-operators-c944h\" (UID: \"d42b609b-dba3-484f-b919-a33fc18e3385\") " pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.880322 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:36 crc kubenswrapper[4894]: I1210 19:01:36.908309 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4lrz"] Dec 10 19:01:37 crc kubenswrapper[4894]: I1210 19:01:37.088762 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4lrz" event={"ID":"6306ae09-754d-4285-8182-4f6433405855","Type":"ContainerStarted","Data":"c53c89a8937170f8ab84b82a516c1f721428339f1e049a3d589df0c67c2860c5"} Dec 10 19:01:37 crc kubenswrapper[4894]: I1210 19:01:37.089060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4lrz" event={"ID":"6306ae09-754d-4285-8182-4f6433405855","Type":"ContainerStarted","Data":"a606b78aa03472b87fd083555d4ccab4cdf6a51c50339972ee260b855e43d477"} Dec 10 19:01:37 crc kubenswrapper[4894]: I1210 19:01:37.106908 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c944h"] Dec 10 19:01:37 crc kubenswrapper[4894]: W1210 19:01:37.185178 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd42b609b_dba3_484f_b919_a33fc18e3385.slice/crio-9959bc76eac2e546682f96ad98e092e03b9fc46b2f5710e1f48f3a5fc2780487 WatchSource:0}: Error finding container 9959bc76eac2e546682f96ad98e092e03b9fc46b2f5710e1f48f3a5fc2780487: Status 404 returned error can't find the container with id 9959bc76eac2e546682f96ad98e092e03b9fc46b2f5710e1f48f3a5fc2780487 Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.095385 4894 generic.go:334] "Generic (PLEG): container finished" podID="6306ae09-754d-4285-8182-4f6433405855" containerID="c53c89a8937170f8ab84b82a516c1f721428339f1e049a3d589df0c67c2860c5" exitCode=0 Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.095429 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4lrz" event={"ID":"6306ae09-754d-4285-8182-4f6433405855","Type":"ContainerDied","Data":"c53c89a8937170f8ab84b82a516c1f721428339f1e049a3d589df0c67c2860c5"} Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.100427 4894 generic.go:334] "Generic (PLEG): container finished" podID="d42b609b-dba3-484f-b919-a33fc18e3385" containerID="35a28e3fdc011a943fad81bc9b8f471e6856630250e391d93aa84c8a80f3027b" exitCode=0 Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.100479 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c944h" event={"ID":"d42b609b-dba3-484f-b919-a33fc18e3385","Type":"ContainerDied","Data":"35a28e3fdc011a943fad81bc9b8f471e6856630250e391d93aa84c8a80f3027b"} Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.100512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c944h" event={"ID":"d42b609b-dba3-484f-b919-a33fc18e3385","Type":"ContainerStarted","Data":"9959bc76eac2e546682f96ad98e092e03b9fc46b2f5710e1f48f3a5fc2780487"} Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.152382 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6zg8t"] Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.155762 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zg8t"] Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.155912 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.158425 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.214680 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8892c15-a6a1-45d4-adf1-84b6e132b411-utilities\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.215136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5vdr\" (UniqueName: \"kubernetes.io/projected/e8892c15-a6a1-45d4-adf1-84b6e132b411-kube-api-access-l5vdr\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.215537 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8892c15-a6a1-45d4-adf1-84b6e132b411-catalog-content\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.317319 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8892c15-a6a1-45d4-adf1-84b6e132b411-utilities\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.317365 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5vdr\" (UniqueName: \"kubernetes.io/projected/e8892c15-a6a1-45d4-adf1-84b6e132b411-kube-api-access-l5vdr\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.317398 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8892c15-a6a1-45d4-adf1-84b6e132b411-catalog-content\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.317795 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8892c15-a6a1-45d4-adf1-84b6e132b411-catalog-content\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.318359 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8892c15-a6a1-45d4-adf1-84b6e132b411-utilities\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.339963 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5vdr\" (UniqueName: \"kubernetes.io/projected/e8892c15-a6a1-45d4-adf1-84b6e132b411-kube-api-access-l5vdr\") pod \"redhat-marketplace-6zg8t\" (UID: \"e8892c15-a6a1-45d4-adf1-84b6e132b411\") " pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.504290 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:38 crc kubenswrapper[4894]: I1210 19:01:38.712448 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zg8t"] Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.108941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4lrz" event={"ID":"6306ae09-754d-4285-8182-4f6433405855","Type":"ContainerStarted","Data":"38fa3ddba19901642cb9b77c89213d541fc2f627c238f13e67b4aac664c96e49"} Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.112810 4894 generic.go:334] "Generic (PLEG): container finished" podID="e8892c15-a6a1-45d4-adf1-84b6e132b411" containerID="08e4b4ed8ea6132beb89745366f4d751e4ff8fb7ebbb58f3c1ab1e62b345eefd" exitCode=0 Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.112882 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zg8t" event={"ID":"e8892c15-a6a1-45d4-adf1-84b6e132b411","Type":"ContainerDied","Data":"08e4b4ed8ea6132beb89745366f4d751e4ff8fb7ebbb58f3c1ab1e62b345eefd"} Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.112909 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zg8t" event={"ID":"e8892c15-a6a1-45d4-adf1-84b6e132b411","Type":"ContainerStarted","Data":"1cc812e5219dcbab74fad736b317afa61308723fbd6aa96f47390f2d4b443506"} Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.146298 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t98mg"] Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.147492 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.149665 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.157429 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t98mg"] Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.227671 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-catalog-content\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.227978 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-utilities\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.228300 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlr7w\" (UniqueName: \"kubernetes.io/projected/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-kube-api-access-vlr7w\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.329750 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-catalog-content\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.329897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-utilities\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.329987 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlr7w\" (UniqueName: \"kubernetes.io/projected/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-kube-api-access-vlr7w\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.331138 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-catalog-content\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.331371 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-utilities\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.352140 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlr7w\" (UniqueName: \"kubernetes.io/projected/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-kube-api-access-vlr7w\") pod \"redhat-operators-t98mg\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.544671 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:39 crc kubenswrapper[4894]: I1210 19:01:39.723284 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t98mg"] Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.121937 4894 generic.go:334] "Generic (PLEG): container finished" podID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerID="6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0" exitCode=0 Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.122022 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerDied","Data":"6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0"} Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.122275 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerStarted","Data":"a254811a8106633f4321a0ec56cd09d42d0e4c86391d1d39f1ea95a44877d40d"} Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.124717 4894 generic.go:334] "Generic (PLEG): container finished" podID="6306ae09-754d-4285-8182-4f6433405855" containerID="38fa3ddba19901642cb9b77c89213d541fc2f627c238f13e67b4aac664c96e49" exitCode=0 Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.124885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4lrz" event={"ID":"6306ae09-754d-4285-8182-4f6433405855","Type":"ContainerDied","Data":"38fa3ddba19901642cb9b77c89213d541fc2f627c238f13e67b4aac664c96e49"} Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.127177 4894 generic.go:334] "Generic (PLEG): container finished" podID="d42b609b-dba3-484f-b919-a33fc18e3385" containerID="16f1b3cb422acea53cae84ae6d5a9bfe664863cda927cd2bb471fb552d5b2083" exitCode=0 Dec 10 19:01:40 crc kubenswrapper[4894]: I1210 19:01:40.127227 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c944h" event={"ID":"d42b609b-dba3-484f-b919-a33fc18e3385","Type":"ContainerDied","Data":"16f1b3cb422acea53cae84ae6d5a9bfe664863cda927cd2bb471fb552d5b2083"} Dec 10 19:01:41 crc kubenswrapper[4894]: I1210 19:01:41.133348 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerStarted","Data":"a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c"} Dec 10 19:01:41 crc kubenswrapper[4894]: I1210 19:01:41.137999 4894 generic.go:334] "Generic (PLEG): container finished" podID="e8892c15-a6a1-45d4-adf1-84b6e132b411" containerID="da2458c7007b241027054e9eb5d248a85beb2a90d453668d9eed4b37ca8e497f" exitCode=0 Dec 10 19:01:41 crc kubenswrapper[4894]: I1210 19:01:41.138096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zg8t" event={"ID":"e8892c15-a6a1-45d4-adf1-84b6e132b411","Type":"ContainerDied","Data":"da2458c7007b241027054e9eb5d248a85beb2a90d453668d9eed4b37ca8e497f"} Dec 10 19:01:41 crc kubenswrapper[4894]: I1210 19:01:41.146971 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c944h" event={"ID":"d42b609b-dba3-484f-b919-a33fc18e3385","Type":"ContainerStarted","Data":"745df0299e0452d2bb532f044238abb995c8d8b17a1cbdce5e21bdca163cf814"} Dec 10 19:01:41 crc kubenswrapper[4894]: I1210 19:01:41.196282 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c944h" podStartSLOduration=2.446144851 podStartE2EDuration="5.196264288s" podCreationTimestamp="2025-12-10 19:01:36 +0000 UTC" firstStartedPulling="2025-12-10 19:01:38.102257649 +0000 UTC m=+380.897105441" lastFinishedPulling="2025-12-10 19:01:40.852377086 +0000 UTC m=+383.647224878" observedRunningTime="2025-12-10 19:01:41.193897314 +0000 UTC m=+383.988745126" watchObservedRunningTime="2025-12-10 19:01:41.196264288 +0000 UTC m=+383.991112080" Dec 10 19:01:42 crc kubenswrapper[4894]: I1210 19:01:42.170911 4894 generic.go:334] "Generic (PLEG): container finished" podID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerID="a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c" exitCode=0 Dec 10 19:01:42 crc kubenswrapper[4894]: I1210 19:01:42.171017 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerDied","Data":"a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c"} Dec 10 19:01:42 crc kubenswrapper[4894]: I1210 19:01:42.173106 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4lrz" event={"ID":"6306ae09-754d-4285-8182-4f6433405855","Type":"ContainerStarted","Data":"909235fbbbab1f34a1a6bce962a38fbde910c6799459445e9e7a84149528fbd4"} Dec 10 19:01:42 crc kubenswrapper[4894]: I1210 19:01:42.218046 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4lrz" podStartSLOduration=3.24885048 podStartE2EDuration="6.218027289s" podCreationTimestamp="2025-12-10 19:01:36 +0000 UTC" firstStartedPulling="2025-12-10 19:01:38.098743965 +0000 UTC m=+380.893591757" lastFinishedPulling="2025-12-10 19:01:41.067920774 +0000 UTC m=+383.862768566" observedRunningTime="2025-12-10 19:01:42.215214043 +0000 UTC m=+385.010061845" watchObservedRunningTime="2025-12-10 19:01:42.218027289 +0000 UTC m=+385.012875091" Dec 10 19:01:42 crc kubenswrapper[4894]: I1210 19:01:42.727412 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" podUID="d03795b4-dfb1-47b7-9cfe-17f03c460f37" containerName="registry" containerID="cri-o://0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69" gracePeriod=30 Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.090455 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.179098 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerStarted","Data":"53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9"} Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.182212 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zg8t" event={"ID":"e8892c15-a6a1-45d4-adf1-84b6e132b411","Type":"ContainerStarted","Data":"a2227067cd43b4c7b366a22f5f217b976394458843cafa619605299f69e495ef"} Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.189783 4894 generic.go:334] "Generic (PLEG): container finished" podID="d03795b4-dfb1-47b7-9cfe-17f03c460f37" containerID="0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69" exitCode=0 Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.189827 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" event={"ID":"d03795b4-dfb1-47b7-9cfe-17f03c460f37","Type":"ContainerDied","Data":"0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69"} Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.189881 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" event={"ID":"d03795b4-dfb1-47b7-9cfe-17f03c460f37","Type":"ContainerDied","Data":"da3c81cce41537d54e696bb310ca0dd32d0b7383a07c6b95afe8dbcf04218d5e"} Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.189900 4894 scope.go:117] "RemoveContainer" containerID="0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.189914 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jnckv" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.199531 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t98mg" podStartSLOduration=1.65368908 podStartE2EDuration="4.199515261s" podCreationTimestamp="2025-12-10 19:01:39 +0000 UTC" firstStartedPulling="2025-12-10 19:01:40.129954756 +0000 UTC m=+382.924802548" lastFinishedPulling="2025-12-10 19:01:42.675780937 +0000 UTC m=+385.470628729" observedRunningTime="2025-12-10 19:01:43.194655431 +0000 UTC m=+385.989503243" watchObservedRunningTime="2025-12-10 19:01:43.199515261 +0000 UTC m=+385.994363053" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.200893 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d03795b4-dfb1-47b7-9cfe-17f03c460f37-ca-trust-extracted\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201037 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d03795b4-dfb1-47b7-9cfe-17f03c460f37-installation-pull-secrets\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201097 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-bound-sa-token\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201346 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201397 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-tls\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201435 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnpsn\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-kube-api-access-lnpsn\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201478 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-trusted-ca\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.201517 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-certificates\") pod \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\" (UID: \"d03795b4-dfb1-47b7-9cfe-17f03c460f37\") " Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.202496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.202542 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.209092 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.210585 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-kube-api-access-lnpsn" (OuterVolumeSpecName: "kube-api-access-lnpsn") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "kube-api-access-lnpsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.210751 4894 scope.go:117] "RemoveContainer" containerID="0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.210913 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.210969 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03795b4-dfb1-47b7-9cfe-17f03c460f37-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: E1210 19:01:43.213235 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69\": container with ID starting with 0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69 not found: ID does not exist" containerID="0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.213315 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69"} err="failed to get container status \"0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69\": rpc error: code = NotFound desc = could not find container \"0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69\": container with ID starting with 0a69ff44cb04040660d07a2e6faf6ef6558f275e169f659f77245c8ed311db69 not found: ID does not exist" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.222908 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d03795b4-dfb1-47b7-9cfe-17f03c460f37-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.224193 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d03795b4-dfb1-47b7-9cfe-17f03c460f37" (UID: "d03795b4-dfb1-47b7-9cfe-17f03c460f37"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.224360 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6zg8t" podStartSLOduration=2.230661381 podStartE2EDuration="5.224303464s" podCreationTimestamp="2025-12-10 19:01:38 +0000 UTC" firstStartedPulling="2025-12-10 19:01:39.114582097 +0000 UTC m=+381.909429919" lastFinishedPulling="2025-12-10 19:01:42.10822421 +0000 UTC m=+384.903072002" observedRunningTime="2025-12-10 19:01:43.216257769 +0000 UTC m=+386.011105561" watchObservedRunningTime="2025-12-10 19:01:43.224303464 +0000 UTC m=+386.019151256" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304035 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304082 4894 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304093 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnpsn\" (UniqueName: \"kubernetes.io/projected/d03795b4-dfb1-47b7-9cfe-17f03c460f37-kube-api-access-lnpsn\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304102 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304112 4894 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d03795b4-dfb1-47b7-9cfe-17f03c460f37-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304121 4894 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d03795b4-dfb1-47b7-9cfe-17f03c460f37-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.304134 4894 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d03795b4-dfb1-47b7-9cfe-17f03c460f37-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.529653 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jnckv"] Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.534283 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jnckv"] Dec 10 19:01:43 crc kubenswrapper[4894]: I1210 19:01:43.549026 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d03795b4-dfb1-47b7-9cfe-17f03c460f37" path="/var/lib/kubelet/pods/d03795b4-dfb1-47b7-9cfe-17f03c460f37/volumes" Dec 10 19:01:46 crc kubenswrapper[4894]: I1210 19:01:46.677379 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:46 crc kubenswrapper[4894]: I1210 19:01:46.677774 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:46 crc kubenswrapper[4894]: I1210 19:01:46.727631 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:46 crc kubenswrapper[4894]: I1210 19:01:46.882023 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:46 crc kubenswrapper[4894]: I1210 19:01:46.882082 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:46 crc kubenswrapper[4894]: I1210 19:01:46.918444 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:47 crc kubenswrapper[4894]: I1210 19:01:47.254359 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c944h" Dec 10 19:01:47 crc kubenswrapper[4894]: I1210 19:01:47.254669 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4lrz" Dec 10 19:01:48 crc kubenswrapper[4894]: I1210 19:01:48.504692 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:48 crc kubenswrapper[4894]: I1210 19:01:48.506049 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:48 crc kubenswrapper[4894]: I1210 19:01:48.553020 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:49 crc kubenswrapper[4894]: I1210 19:01:49.293353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6zg8t" Dec 10 19:01:49 crc kubenswrapper[4894]: I1210 19:01:49.572702 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:49 crc kubenswrapper[4894]: I1210 19:01:49.572774 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:49 crc kubenswrapper[4894]: I1210 19:01:49.576932 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:01:49 crc kubenswrapper[4894]: I1210 19:01:49.577029 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:01:49 crc kubenswrapper[4894]: I1210 19:01:49.615621 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:01:50 crc kubenswrapper[4894]: I1210 19:01:50.283332 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:02:19 crc kubenswrapper[4894]: I1210 19:02:19.577002 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:02:19 crc kubenswrapper[4894]: I1210 19:02:19.577661 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:02:19 crc kubenswrapper[4894]: I1210 19:02:19.577724 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:02:19 crc kubenswrapper[4894]: I1210 19:02:19.578642 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a67b89ac2990b2f668b7964537a50bb5050a9050493de8896a0931aefdc247e9"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:02:19 crc kubenswrapper[4894]: I1210 19:02:19.578745 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://a67b89ac2990b2f668b7964537a50bb5050a9050493de8896a0931aefdc247e9" gracePeriod=600 Dec 10 19:02:20 crc kubenswrapper[4894]: I1210 19:02:20.415395 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="a67b89ac2990b2f668b7964537a50bb5050a9050493de8896a0931aefdc247e9" exitCode=0 Dec 10 19:02:20 crc kubenswrapper[4894]: I1210 19:02:20.415448 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"a67b89ac2990b2f668b7964537a50bb5050a9050493de8896a0931aefdc247e9"} Dec 10 19:02:20 crc kubenswrapper[4894]: I1210 19:02:20.415747 4894 scope.go:117] "RemoveContainer" containerID="f772dfa28288ace770e5e2bbb594ae764e82ae234a7b95ffda29cf9f4dc59955" Dec 10 19:02:21 crc kubenswrapper[4894]: I1210 19:02:21.425451 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"1143033875ce2861ff6783d3b0a8fe27cde229985e69e12e38acac4bb01e1718"} Dec 10 19:04:17 crc kubenswrapper[4894]: I1210 19:04:17.793355 4894 scope.go:117] "RemoveContainer" containerID="680b030bfd1523df05bb4649b20e1b41efc2252370e3ae33eb4ea034972c813b" Dec 10 19:04:49 crc kubenswrapper[4894]: I1210 19:04:49.577629 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:04:49 crc kubenswrapper[4894]: I1210 19:04:49.578667 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:05:17 crc kubenswrapper[4894]: I1210 19:05:17.831385 4894 scope.go:117] "RemoveContainer" containerID="db0aa1826da1f98708f7e6aee7db016fe6fcb7cc710455613cf3bc808c1ae3de" Dec 10 19:05:17 crc kubenswrapper[4894]: I1210 19:05:17.863248 4894 scope.go:117] "RemoveContainer" containerID="78455d4420edbfeb783b49166aa9dea7eb2387c74fbda2dce34818bdd6b7ad1d" Dec 10 19:05:19 crc kubenswrapper[4894]: I1210 19:05:19.577301 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:05:19 crc kubenswrapper[4894]: I1210 19:05:19.577373 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.577607 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.579996 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.580158 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.580923 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1143033875ce2861ff6783d3b0a8fe27cde229985e69e12e38acac4bb01e1718"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.581109 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://1143033875ce2861ff6783d3b0a8fe27cde229985e69e12e38acac4bb01e1718" gracePeriod=600 Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.878050 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="1143033875ce2861ff6783d3b0a8fe27cde229985e69e12e38acac4bb01e1718" exitCode=0 Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.878147 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"1143033875ce2861ff6783d3b0a8fe27cde229985e69e12e38acac4bb01e1718"} Dec 10 19:05:49 crc kubenswrapper[4894]: I1210 19:05:49.878598 4894 scope.go:117] "RemoveContainer" containerID="a67b89ac2990b2f668b7964537a50bb5050a9050493de8896a0931aefdc247e9" Dec 10 19:05:50 crc kubenswrapper[4894]: I1210 19:05:50.901695 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"4ec2353aa9d8d229a346f66ddc339bc5b86fce1f5b3bf7d51741a93c0c96e096"} Dec 10 19:07:04 crc kubenswrapper[4894]: I1210 19:07:04.940192 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm"] Dec 10 19:07:04 crc kubenswrapper[4894]: E1210 19:07:04.940990 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03795b4-dfb1-47b7-9cfe-17f03c460f37" containerName="registry" Dec 10 19:07:04 crc kubenswrapper[4894]: I1210 19:07:04.941005 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03795b4-dfb1-47b7-9cfe-17f03c460f37" containerName="registry" Dec 10 19:07:04 crc kubenswrapper[4894]: I1210 19:07:04.941116 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03795b4-dfb1-47b7-9cfe-17f03c460f37" containerName="registry" Dec 10 19:07:04 crc kubenswrapper[4894]: I1210 19:07:04.941898 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:04 crc kubenswrapper[4894]: I1210 19:07:04.943371 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:07:04 crc kubenswrapper[4894]: I1210 19:07:04.950282 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm"] Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.052746 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.052909 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn928\" (UniqueName: \"kubernetes.io/projected/d191af44-bc04-4529-93a4-b58a773920f5-kube-api-access-qn928\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.052945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.156138 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn928\" (UniqueName: \"kubernetes.io/projected/d191af44-bc04-4529-93a4-b58a773920f5-kube-api-access-qn928\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.156218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.156329 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.156765 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.157000 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.191092 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn928\" (UniqueName: \"kubernetes.io/projected/d191af44-bc04-4529-93a4-b58a773920f5-kube-api-access-qn928\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.370507 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:05 crc kubenswrapper[4894]: I1210 19:07:05.635541 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm"] Dec 10 19:07:06 crc kubenswrapper[4894]: I1210 19:07:06.374921 4894 generic.go:334] "Generic (PLEG): container finished" podID="d191af44-bc04-4529-93a4-b58a773920f5" containerID="06df5353e46464d7ea4c6669146b938edaf7afb025ec98be64a22331b05edcd4" exitCode=0 Dec 10 19:07:06 crc kubenswrapper[4894]: I1210 19:07:06.374965 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" event={"ID":"d191af44-bc04-4529-93a4-b58a773920f5","Type":"ContainerDied","Data":"06df5353e46464d7ea4c6669146b938edaf7afb025ec98be64a22331b05edcd4"} Dec 10 19:07:06 crc kubenswrapper[4894]: I1210 19:07:06.375159 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" event={"ID":"d191af44-bc04-4529-93a4-b58a773920f5","Type":"ContainerStarted","Data":"367214cbb11da19fa185796cd05fcb99b28fcfd20344ac261d9ef7aaa187cfc2"} Dec 10 19:07:06 crc kubenswrapper[4894]: I1210 19:07:06.376961 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:07:08 crc kubenswrapper[4894]: I1210 19:07:08.386200 4894 generic.go:334] "Generic (PLEG): container finished" podID="d191af44-bc04-4529-93a4-b58a773920f5" containerID="6bd81054047c4a91b84b8fc623c459d00961e5102fcd09a043a4832f79672014" exitCode=0 Dec 10 19:07:08 crc kubenswrapper[4894]: I1210 19:07:08.386271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" event={"ID":"d191af44-bc04-4529-93a4-b58a773920f5","Type":"ContainerDied","Data":"6bd81054047c4a91b84b8fc623c459d00961e5102fcd09a043a4832f79672014"} Dec 10 19:07:09 crc kubenswrapper[4894]: I1210 19:07:09.395479 4894 generic.go:334] "Generic (PLEG): container finished" podID="d191af44-bc04-4529-93a4-b58a773920f5" containerID="01af81ceb9052025b49c191b0f083b9a509afc60f365412ca73a51472275a43a" exitCode=0 Dec 10 19:07:09 crc kubenswrapper[4894]: I1210 19:07:09.395532 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" event={"ID":"d191af44-bc04-4529-93a4-b58a773920f5","Type":"ContainerDied","Data":"01af81ceb9052025b49c191b0f083b9a509afc60f365412ca73a51472275a43a"} Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.729178 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.831770 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-bundle\") pod \"d191af44-bc04-4529-93a4-b58a773920f5\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.831832 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-util\") pod \"d191af44-bc04-4529-93a4-b58a773920f5\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.831924 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn928\" (UniqueName: \"kubernetes.io/projected/d191af44-bc04-4529-93a4-b58a773920f5-kube-api-access-qn928\") pod \"d191af44-bc04-4529-93a4-b58a773920f5\" (UID: \"d191af44-bc04-4529-93a4-b58a773920f5\") " Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.836711 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-bundle" (OuterVolumeSpecName: "bundle") pod "d191af44-bc04-4529-93a4-b58a773920f5" (UID: "d191af44-bc04-4529-93a4-b58a773920f5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.837823 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d191af44-bc04-4529-93a4-b58a773920f5-kube-api-access-qn928" (OuterVolumeSpecName: "kube-api-access-qn928") pod "d191af44-bc04-4529-93a4-b58a773920f5" (UID: "d191af44-bc04-4529-93a4-b58a773920f5"). InnerVolumeSpecName "kube-api-access-qn928". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.851734 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-util" (OuterVolumeSpecName: "util") pod "d191af44-bc04-4529-93a4-b58a773920f5" (UID: "d191af44-bc04-4529-93a4-b58a773920f5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.933406 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.933454 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d191af44-bc04-4529-93a4-b58a773920f5-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:10 crc kubenswrapper[4894]: I1210 19:07:10.933472 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn928\" (UniqueName: \"kubernetes.io/projected/d191af44-bc04-4529-93a4-b58a773920f5-kube-api-access-qn928\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:11 crc kubenswrapper[4894]: I1210 19:07:11.414366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" event={"ID":"d191af44-bc04-4529-93a4-b58a773920f5","Type":"ContainerDied","Data":"367214cbb11da19fa185796cd05fcb99b28fcfd20344ac261d9ef7aaa187cfc2"} Dec 10 19:07:11 crc kubenswrapper[4894]: I1210 19:07:11.414437 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367214cbb11da19fa185796cd05fcb99b28fcfd20344ac261d9ef7aaa187cfc2" Dec 10 19:07:11 crc kubenswrapper[4894]: I1210 19:07:11.414459 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130100 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jqwjd"] Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130668 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-controller" containerID="cri-o://52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130720 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="nbdb" containerID="cri-o://3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130745 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130775 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="northd" containerID="cri-o://8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130778 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-node" containerID="cri-o://024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130823 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="sbdb" containerID="cri-o://fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.130793 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-acl-logging" containerID="cri-o://feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.189811 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" containerID="cri-o://8e5d9ff95576272fc324ba39a9ab49fcad475765ccacc2d6184b5302bbbf7ff8" gracePeriod=30 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.473201 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovnkube-controller/3.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.475450 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovn-acl-logging/0.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.475920 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovn-controller/0.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476198 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="8e5d9ff95576272fc324ba39a9ab49fcad475765ccacc2d6184b5302bbbf7ff8" exitCode=0 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476282 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb" exitCode=0 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476337 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9" exitCode=0 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476383 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e" exitCode=0 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476451 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2" exitCode=143 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476531 4894 generic.go:334] "Generic (PLEG): container finished" podID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerID="52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702" exitCode=143 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476276 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"8e5d9ff95576272fc324ba39a9ab49fcad475765ccacc2d6184b5302bbbf7ff8"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476634 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476648 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476657 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476676 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.476692 4894 scope.go:117] "RemoveContainer" containerID="e3bbb4919314a434e1b26a22bb39945eaabfdf5a5879577e304092a0e93e7e88" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.483692 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/2.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.484390 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/1.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.484476 4894 generic.go:334] "Generic (PLEG): container finished" podID="d9392e0e-ca87-437a-8ba1-9c5d34a50660" containerID="d422371c764f725f4e201df30c379baf3507de97186b2eeae087e857e84a3523" exitCode=2 Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.484580 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerDied","Data":"d422371c764f725f4e201df30c379baf3507de97186b2eeae087e857e84a3523"} Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.485407 4894 scope.go:117] "RemoveContainer" containerID="d422371c764f725f4e201df30c379baf3507de97186b2eeae087e857e84a3523" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.485611 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-msfl4_openshift-multus(d9392e0e-ca87-437a-8ba1-9c5d34a50660)\"" pod="openshift-multus/multus-msfl4" podUID="d9392e0e-ca87-437a-8ba1-9c5d34a50660" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.504880 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovn-acl-logging/0.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.513581 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jqwjd_8de31cb9-e20f-4578-b0ac-b03df0bc133c/ovn-controller/0.log" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.515597 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.519371 4894 scope.go:117] "RemoveContainer" containerID="dc57a1387f09e35821d6f3f412f21cba558965a6f1b73381668b99377f53da53" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617785 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-systemd\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617819 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-slash\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617873 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617903 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-node-log\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617932 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-netd\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617948 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-var-lib-openvswitch\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617940 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-slash" (OuterVolumeSpecName: "host-slash") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617980 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-config\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.617998 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-netns\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618001 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-node-log" (OuterVolumeSpecName: "node-log") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618019 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovn-node-metrics-cert\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618025 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618043 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-script-lib\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618062 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-ovn-kubernetes\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618084 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnqbs\" (UniqueName: \"kubernetes.io/projected/8de31cb9-e20f-4578-b0ac-b03df0bc133c-kube-api-access-bnqbs\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618112 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-etc-openvswitch\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618132 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-systemd-units\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618149 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-log-socket\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618165 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-bin\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618189 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-openvswitch\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618207 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-kubelet\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618255 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-ovn\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618271 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-env-overrides\") pod \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\" (UID: \"8de31cb9-e20f-4578-b0ac-b03df0bc133c\") " Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618352 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618399 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618421 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618440 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-log-socket" (OuterVolumeSpecName: "log-socket") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618461 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618491 4894 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-slash\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618499 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618504 4894 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618515 4894 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-node-log\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618525 4894 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618533 4894 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618541 4894 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-log-socket\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618549 4894 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618521 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618561 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618540 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.618978 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.619129 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.619152 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.619218 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.624511 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8de31cb9-e20f-4578-b0ac-b03df0bc133c-kube-api-access-bnqbs" (OuterVolumeSpecName: "kube-api-access-bnqbs") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "kube-api-access-bnqbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.628612 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.632261 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "8de31cb9-e20f-4578-b0ac-b03df0bc133c" (UID: "8de31cb9-e20f-4578-b0ac-b03df0bc133c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.639714 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rc82k"] Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.639918 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.639929 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.639940 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="util" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.639945 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="util" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.639955 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="northd" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.639960 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="northd" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.639968 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.639974 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.639980 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.639987 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.639998 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="nbdb" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640005 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="nbdb" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640017 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-node" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640024 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-node" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640035 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640043 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640053 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-acl-logging" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640061 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-acl-logging" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640070 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640078 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640089 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="extract" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640097 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="extract" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640106 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kubecfg-setup" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640114 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kubecfg-setup" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640124 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="pull" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640129 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="pull" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640139 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="sbdb" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640144 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="sbdb" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640228 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="northd" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640238 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640245 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640254 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640260 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="sbdb" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640266 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d191af44-bc04-4529-93a4-b58a773920f5" containerName="extract" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640272 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640278 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="nbdb" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640288 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-acl-logging" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640293 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovn-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640300 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="kube-rbac-proxy-node" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640383 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640391 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: E1210 19:07:17.640398 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640403 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640500 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.640512 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" containerName="ovnkube-controller" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.642276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-log-socket\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-slash\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-run-ovn-kubernetes\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720125 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720175 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovnkube-config\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720207 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-ovn\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720231 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmzrc\" (UniqueName: \"kubernetes.io/projected/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-kube-api-access-vmzrc\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720247 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-node-log\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720266 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-kubelet\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720282 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-cni-bin\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720299 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-systemd\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720313 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-systemd-units\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovnkube-script-lib\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720342 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-cni-netd\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720358 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-var-lib-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovn-node-metrics-cert\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-run-netns\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720416 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-env-overrides\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720439 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720457 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-etc-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720491 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720502 4894 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720511 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720519 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720528 4894 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720537 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnqbs\" (UniqueName: \"kubernetes.io/projected/8de31cb9-e20f-4578-b0ac-b03df0bc133c-kube-api-access-bnqbs\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720545 4894 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720553 4894 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720561 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8de31cb9-e20f-4578-b0ac-b03df0bc133c-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720569 4894 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720577 4894 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720586 4894 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.720594 4894 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8de31cb9-e20f-4578-b0ac-b03df0bc133c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821142 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-slash\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821183 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-log-socket\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-run-ovn-kubernetes\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821229 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821252 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovnkube-config\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821275 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-ovn\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821279 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-slash\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821295 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-log-socket\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-run-ovn-kubernetes\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmzrc\" (UniqueName: \"kubernetes.io/projected/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-kube-api-access-vmzrc\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821380 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-ovn\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821391 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-node-log\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821419 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821423 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-kubelet\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-kubelet\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821464 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-node-log\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-cni-bin\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821499 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-systemd\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821523 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-systemd-units\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821545 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovnkube-script-lib\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821553 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-cni-bin\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821567 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-cni-netd\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-systemd-units\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821594 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-var-lib-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821601 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-run-systemd\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovn-node-metrics-cert\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821625 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-cni-netd\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-run-netns\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-env-overrides\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821744 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821760 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-etc-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821833 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-etc-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821868 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-run-netns\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821916 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-var-lib-openvswitch\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.821963 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.822268 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovnkube-config\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.822289 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-env-overrides\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.822459 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovnkube-script-lib\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.828311 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-ovn-node-metrics-cert\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.839554 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmzrc\" (UniqueName: \"kubernetes.io/projected/e1d18047-ca78-4f09-9ae0-a1d0b2f06a04-kube-api-access-vmzrc\") pod \"ovnkube-node-rc82k\" (UID: \"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04\") " pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.917888 4894 scope.go:117] "RemoveContainer" containerID="fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.931945 4894 scope.go:117] "RemoveContainer" containerID="024fffde866dc696cf97182a23306904f7e2327e758425daf86c484f92d55a9e" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.946176 4894 scope.go:117] "RemoveContainer" containerID="8e5d9ff95576272fc324ba39a9ab49fcad475765ccacc2d6184b5302bbbf7ff8" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.956443 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.963975 4894 scope.go:117] "RemoveContainer" containerID="3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292" Dec 10 19:07:17 crc kubenswrapper[4894]: I1210 19:07:17.977584 4894 scope.go:117] "RemoveContainer" containerID="feeb72256f4d6a0b5e2bda8a87c09a83f94c2d9ceeb5819be543daf373bd32f2" Dec 10 19:07:18 crc kubenswrapper[4894]: W1210 19:07:18.000059 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1d18047_ca78_4f09_9ae0_a1d0b2f06a04.slice/crio-d711bc82508804d0f70ad488f399ca5e16b4d3c148dae60349e2ce50f3fd75e7 WatchSource:0}: Error finding container d711bc82508804d0f70ad488f399ca5e16b4d3c148dae60349e2ce50f3fd75e7: Status 404 returned error can't find the container with id d711bc82508804d0f70ad488f399ca5e16b4d3c148dae60349e2ce50f3fd75e7 Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.006807 4894 scope.go:117] "RemoveContainer" containerID="52fe1af73bb16ff0a2bfc6f7be8c7d039ccfacabbae1ac774b0b8c19dc933702" Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.025636 4894 scope.go:117] "RemoveContainer" containerID="8a65165affb133291357f49dad5ec0c803017165ec7366da7745d608bf4090bb" Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.043498 4894 scope.go:117] "RemoveContainer" containerID="add1ab814931d5f727185e08e6509adb7bd9973b78ec23574bf59c251d00e7e9" Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.062329 4894 scope.go:117] "RemoveContainer" containerID="7895c069f9de1e14f7cd1166fb0e9b79dddaa56e96092183fceaee9bdf11c4e4" Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.490538 4894 generic.go:334] "Generic (PLEG): container finished" podID="e1d18047-ca78-4f09-9ae0-a1d0b2f06a04" containerID="a88617afb827c82ac4e12e8071d32c5ea6ddf53b3dceb5546281d673486b9141" exitCode=0 Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.490611 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerDied","Data":"a88617afb827c82ac4e12e8071d32c5ea6ddf53b3dceb5546281d673486b9141"} Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.490656 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"d711bc82508804d0f70ad488f399ca5e16b4d3c148dae60349e2ce50f3fd75e7"} Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.493448 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/2.log" Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.493544 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.493552 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"fe165ac7c20715dcd92d8328136c897b4cda07f103372af94f8210f15d15a0b2"} Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.493629 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"3576073c700fa98529088ca7158fbab678c1257832b458a8400f550406009292"} Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.493656 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jqwjd" event={"ID":"8de31cb9-e20f-4578-b0ac-b03df0bc133c","Type":"ContainerDied","Data":"760360a48f31b9ed7a20ab001b50f3b54e7cfe78684b3dc94edc35a1d3eac9f4"} Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.554810 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jqwjd"] Dec 10 19:07:18 crc kubenswrapper[4894]: I1210 19:07:18.555380 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jqwjd"] Dec 10 19:07:19 crc kubenswrapper[4894]: I1210 19:07:19.501205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"9ce44ad5afc5be2c8c6267241d6252a37a406f14e98e8e72a6a26649b4b195a8"} Dec 10 19:07:19 crc kubenswrapper[4894]: I1210 19:07:19.501537 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"c047a883823dc33b5e25d633813990a18d8a9fa8489a8884a922a7d0adcaa651"} Dec 10 19:07:19 crc kubenswrapper[4894]: I1210 19:07:19.501548 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"d26fbfee327265f7d964323ce4ba6cef9abf70c8e36f6bb7c6201f85791cf4ce"} Dec 10 19:07:19 crc kubenswrapper[4894]: I1210 19:07:19.501556 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"cb444d2deefb0509e4e5b7b63e534206664529ced05031e726106a1750e99d8d"} Dec 10 19:07:19 crc kubenswrapper[4894]: I1210 19:07:19.501565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"003f6c8329438c7beee7842c98c2032afdb69e928d2875d5d040787267f3f7ca"} Dec 10 19:07:19 crc kubenswrapper[4894]: I1210 19:07:19.552084 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8de31cb9-e20f-4578-b0ac-b03df0bc133c" path="/var/lib/kubelet/pods/8de31cb9-e20f-4578-b0ac-b03df0bc133c/volumes" Dec 10 19:07:20 crc kubenswrapper[4894]: I1210 19:07:20.510498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"ca1d38a7f27029140328e44d712d11651d7e4029bf254aebadcfed621e32dc33"} Dec 10 19:07:22 crc kubenswrapper[4894]: I1210 19:07:22.523558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"274b9e19acb09f2f7e402f50127ecf8fd53bd12b0d1a2ddd054b109dde1f4b2b"} Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.573508 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g"] Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.574137 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:23 crc kubenswrapper[4894]: W1210 19:07:23.576580 4894 reflector.go:561] object-"openshift-operators"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Dec 10 19:07:23 crc kubenswrapper[4894]: E1210 19:07:23.578518 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.577601 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-w88js" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.578788 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.700315 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj"] Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.701104 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.703465 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.708059 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk"] Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.708832 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.710302 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzbbl\" (UniqueName: \"kubernetes.io/projected/3d09cb3e-2427-4282-8112-806df77eb803-kube-api-access-zzbbl\") pod \"obo-prometheus-operator-668cf9dfbb-z6v6g\" (UID: \"3d09cb3e-2427-4282-8112-806df77eb803\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.712559 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-8fnpw" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.806116 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-89qkl"] Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.807167 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.809094 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.811558 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f23b2be9-d333-433e-8381-baec276a1b30-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj\" (UID: \"f23b2be9-d333-433e-8381-baec276a1b30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.811679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1b35687b-c053-4902-8ae4-146c82ae486c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk\" (UID: \"1b35687b-c053-4902-8ae4-146c82ae486c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.811715 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzbbl\" (UniqueName: \"kubernetes.io/projected/3d09cb3e-2427-4282-8112-806df77eb803-kube-api-access-zzbbl\") pod \"obo-prometheus-operator-668cf9dfbb-z6v6g\" (UID: \"3d09cb3e-2427-4282-8112-806df77eb803\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.811764 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1b35687b-c053-4902-8ae4-146c82ae486c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk\" (UID: \"1b35687b-c053-4902-8ae4-146c82ae486c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.811792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f23b2be9-d333-433e-8381-baec276a1b30-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj\" (UID: \"f23b2be9-d333-433e-8381-baec276a1b30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.812650 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-44mfl" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.903734 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-67pkr"] Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.904557 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.907017 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-kx7n6" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.912733 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1b35687b-c053-4902-8ae4-146c82ae486c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk\" (UID: \"1b35687b-c053-4902-8ae4-146c82ae486c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.912820 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1b35687b-c053-4902-8ae4-146c82ae486c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk\" (UID: \"1b35687b-c053-4902-8ae4-146c82ae486c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.912901 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f23b2be9-d333-433e-8381-baec276a1b30-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj\" (UID: \"f23b2be9-d333-433e-8381-baec276a1b30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.912963 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f23b2be9-d333-433e-8381-baec276a1b30-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj\" (UID: \"f23b2be9-d333-433e-8381-baec276a1b30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.912991 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.913049 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq89v\" (UniqueName: \"kubernetes.io/projected/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-kube-api-access-qq89v\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.918255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f23b2be9-d333-433e-8381-baec276a1b30-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj\" (UID: \"f23b2be9-d333-433e-8381-baec276a1b30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.918279 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1b35687b-c053-4902-8ae4-146c82ae486c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk\" (UID: \"1b35687b-c053-4902-8ae4-146c82ae486c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.919054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f23b2be9-d333-433e-8381-baec276a1b30-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj\" (UID: \"f23b2be9-d333-433e-8381-baec276a1b30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:23 crc kubenswrapper[4894]: I1210 19:07:23.919059 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1b35687b-c053-4902-8ae4-146c82ae486c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk\" (UID: \"1b35687b-c053-4902-8ae4-146c82ae486c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.013627 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.013674 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c6lm\" (UniqueName: \"kubernetes.io/projected/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-kube-api-access-6c6lm\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.013716 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq89v\" (UniqueName: \"kubernetes.io/projected/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-kube-api-access-qq89v\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.013764 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-openshift-service-ca\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.019565 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.023025 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.034015 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.058963 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(d8a3f8f9afa5d12c41e5499cf79009aefdccc2bef01fbbecce7311cf1bf56c0a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.059028 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(d8a3f8f9afa5d12c41e5499cf79009aefdccc2bef01fbbecce7311cf1bf56c0a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.059048 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(d8a3f8f9afa5d12c41e5499cf79009aefdccc2bef01fbbecce7311cf1bf56c0a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.059088 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators(f23b2be9-d333-433e-8381-baec276a1b30)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators(f23b2be9-d333-433e-8381-baec276a1b30)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(d8a3f8f9afa5d12c41e5499cf79009aefdccc2bef01fbbecce7311cf1bf56c0a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" podUID="f23b2be9-d333-433e-8381-baec276a1b30" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.063248 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(bea467ebb24b436dae7416bc6e0c6aedd22fddf7fb547bb75dbe84ec0752e2b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.063278 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(bea467ebb24b436dae7416bc6e0c6aedd22fddf7fb547bb75dbe84ec0752e2b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.063293 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(bea467ebb24b436dae7416bc6e0c6aedd22fddf7fb547bb75dbe84ec0752e2b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.063317 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators(1b35687b-c053-4902-8ae4-146c82ae486c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators(1b35687b-c053-4902-8ae4-146c82ae486c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(bea467ebb24b436dae7416bc6e0c6aedd22fddf7fb547bb75dbe84ec0752e2b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" podUID="1b35687b-c053-4902-8ae4-146c82ae486c" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.115027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c6lm\" (UniqueName: \"kubernetes.io/projected/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-kube-api-access-6c6lm\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.115135 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-openshift-service-ca\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.115854 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-openshift-service-ca\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.560065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" event={"ID":"e1d18047-ca78-4f09-9ae0-a1d0b2f06a04","Type":"ContainerStarted","Data":"8069bd2c4122b8aa019c83dcf0600398f0fb06562ef7f60bca6c9f8a3830475d"} Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.560976 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.561193 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.561436 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.596050 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.597136 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:24 crc kubenswrapper[4894]: I1210 19:07:24.601504 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" podStartSLOduration=7.6014887810000005 podStartE2EDuration="7.601488781s" podCreationTimestamp="2025-12-10 19:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:07:24.599925119 +0000 UTC m=+727.394772921" watchObservedRunningTime="2025-12-10 19:07:24.601488781 +0000 UTC m=+727.396336583" Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.832748 4894 projected.go:288] Couldn't get configMap openshift-operators/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.832813 4894 projected.go:194] Error preparing data for projected volume kube-api-access-zzbbl for pod openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g: failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:24 crc kubenswrapper[4894]: E1210 19:07:24.832897 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3d09cb3e-2427-4282-8112-806df77eb803-kube-api-access-zzbbl podName:3d09cb3e-2427-4282-8112-806df77eb803 nodeName:}" failed. No retries permitted until 2025-12-10 19:07:25.332876878 +0000 UTC m=+728.127724670 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-zzbbl" (UniqueName: "kubernetes.io/projected/3d09cb3e-2427-4282-8112-806df77eb803-kube-api-access-zzbbl") pod "obo-prometheus-operator-668cf9dfbb-z6v6g" (UID: "3d09cb3e-2427-4282-8112-806df77eb803") : failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.030780 4894 projected.go:288] Couldn't get configMap openshift-operators/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.030861 4894 projected.go:194] Error preparing data for projected volume kube-api-access-qq89v for pod openshift-operators/observability-operator-d8bb48f5d-89qkl: failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.030932 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-kube-api-access-qq89v podName:641b65b6-0781-479e-ac89-c1cdeb2e5b2d nodeName:}" failed. No retries permitted until 2025-12-10 19:07:25.530910954 +0000 UTC m=+728.325758746 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qq89v" (UniqueName: "kubernetes.io/projected/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-kube-api-access-qq89v") pod "observability-operator-d8bb48f5d-89qkl" (UID: "641b65b6-0781-479e-ac89-c1cdeb2e5b2d") : failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.129533 4894 projected.go:288] Couldn't get configMap openshift-operators/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.129599 4894 projected.go:194] Error preparing data for projected volume kube-api-access-6c6lm for pod openshift-operators/perses-operator-5446b9c989-67pkr: failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.129683 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-kube-api-access-6c6lm podName:91e1df09-581a-4dc8-9dae-6834c2bb3bfe nodeName:}" failed. No retries permitted until 2025-12-10 19:07:25.629652753 +0000 UTC m=+728.424500615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6c6lm" (UniqueName: "kubernetes.io/projected/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-kube-api-access-6c6lm") pod "perses-operator-5446b9c989-67pkr" (UID: "91e1df09-581a-4dc8-9dae-6834c2bb3bfe") : failed to sync configmap cache: timed out waiting for the condition Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.148959 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.234891 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk"] Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.235049 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.236116 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.238064 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g"] Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.241195 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-89qkl"] Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.246818 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-67pkr"] Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.249605 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj"] Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.249712 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.250111 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.264122 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(35919a7b1267796e816d88ebbe68f0567bfb1bbc142af1d395293a4e25c73fb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.264197 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(35919a7b1267796e816d88ebbe68f0567bfb1bbc142af1d395293a4e25c73fb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.264216 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(35919a7b1267796e816d88ebbe68f0567bfb1bbc142af1d395293a4e25c73fb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.264262 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators(1b35687b-c053-4902-8ae4-146c82ae486c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators(1b35687b-c053-4902-8ae4-146c82ae486c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(35919a7b1267796e816d88ebbe68f0567bfb1bbc142af1d395293a4e25c73fb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" podUID="1b35687b-c053-4902-8ae4-146c82ae486c" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.276148 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(441baa79738eb3634e1c338d8e828121b75c5c4207d759378110bf7a40a38313): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.276213 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(441baa79738eb3634e1c338d8e828121b75c5c4207d759378110bf7a40a38313): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.276232 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(441baa79738eb3634e1c338d8e828121b75c5c4207d759378110bf7a40a38313): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.276278 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators(f23b2be9-d333-433e-8381-baec276a1b30)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators(f23b2be9-d333-433e-8381-baec276a1b30)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(441baa79738eb3634e1c338d8e828121b75c5c4207d759378110bf7a40a38313): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" podUID="f23b2be9-d333-433e-8381-baec276a1b30" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.429897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzbbl\" (UniqueName: \"kubernetes.io/projected/3d09cb3e-2427-4282-8112-806df77eb803-kube-api-access-zzbbl\") pod \"obo-prometheus-operator-668cf9dfbb-z6v6g\" (UID: \"3d09cb3e-2427-4282-8112-806df77eb803\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.434062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzbbl\" (UniqueName: \"kubernetes.io/projected/3d09cb3e-2427-4282-8112-806df77eb803-kube-api-access-zzbbl\") pod \"obo-prometheus-operator-668cf9dfbb-z6v6g\" (UID: \"3d09cb3e-2427-4282-8112-806df77eb803\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.531308 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq89v\" (UniqueName: \"kubernetes.io/projected/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-kube-api-access-qq89v\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.534756 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq89v\" (UniqueName: \"kubernetes.io/projected/641b65b6-0781-479e-ac89-c1cdeb2e5b2d-kube-api-access-qq89v\") pod \"observability-operator-d8bb48f5d-89qkl\" (UID: \"641b65b6-0781-479e-ac89-c1cdeb2e5b2d\") " pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.626556 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.637691 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c6lm\" (UniqueName: \"kubernetes.io/projected/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-kube-api-access-6c6lm\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.642432 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c6lm\" (UniqueName: \"kubernetes.io/projected/91e1df09-581a-4dc8-9dae-6834c2bb3bfe-kube-api-access-6c6lm\") pod \"perses-operator-5446b9c989-67pkr\" (UID: \"91e1df09-581a-4dc8-9dae-6834c2bb3bfe\") " pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.664468 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1b5aa7f2783cab0e1f774b084d90c6ff52a709a5f389c8080680c6e1c25fa6da): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.664549 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1b5aa7f2783cab0e1f774b084d90c6ff52a709a5f389c8080680c6e1c25fa6da): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.664586 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1b5aa7f2783cab0e1f774b084d90c6ff52a709a5f389c8080680c6e1c25fa6da): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.664671 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-89qkl_openshift-operators(641b65b6-0781-479e-ac89-c1cdeb2e5b2d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-89qkl_openshift-operators(641b65b6-0781-479e-ac89-c1cdeb2e5b2d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1b5aa7f2783cab0e1f774b084d90c6ff52a709a5f389c8080680c6e1c25fa6da): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" podUID="641b65b6-0781-479e-ac89-c1cdeb2e5b2d" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.689238 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.714113 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(a57842b87fe5c9cd2a8ad3227144e15b0bfe6b18f2409bd0ac90656f2e99dfd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.714226 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(a57842b87fe5c9cd2a8ad3227144e15b0bfe6b18f2409bd0ac90656f2e99dfd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.714261 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(a57842b87fe5c9cd2a8ad3227144e15b0bfe6b18f2409bd0ac90656f2e99dfd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.714333 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators(3d09cb3e-2427-4282-8112-806df77eb803)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators(3d09cb3e-2427-4282-8112-806df77eb803)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(a57842b87fe5c9cd2a8ad3227144e15b0bfe6b18f2409bd0ac90656f2e99dfd2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" podUID="3d09cb3e-2427-4282-8112-806df77eb803" Dec 10 19:07:25 crc kubenswrapper[4894]: I1210 19:07:25.797024 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.820493 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(9d8fc841106ebcf8611af58af838dee4226e60479a1b4d9986d6e5702c008e84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.820576 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(9d8fc841106ebcf8611af58af838dee4226e60479a1b4d9986d6e5702c008e84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.820612 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(9d8fc841106ebcf8611af58af838dee4226e60479a1b4d9986d6e5702c008e84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:25 crc kubenswrapper[4894]: E1210 19:07:25.820672 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-67pkr_openshift-operators(91e1df09-581a-4dc8-9dae-6834c2bb3bfe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-67pkr_openshift-operators(91e1df09-581a-4dc8-9dae-6834c2bb3bfe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(9d8fc841106ebcf8611af58af838dee4226e60479a1b4d9986d6e5702c008e84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-67pkr" podUID="91e1df09-581a-4dc8-9dae-6834c2bb3bfe" Dec 10 19:07:26 crc kubenswrapper[4894]: I1210 19:07:26.569522 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:26 crc kubenswrapper[4894]: I1210 19:07:26.569545 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:26 crc kubenswrapper[4894]: I1210 19:07:26.569791 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:26 crc kubenswrapper[4894]: I1210 19:07:26.570340 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:26 crc kubenswrapper[4894]: I1210 19:07:26.570357 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:26 crc kubenswrapper[4894]: I1210 19:07:26.570489 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.604203 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(1411909703f516361a43617f788315863cba8f0ff9593367629772219adf5dbc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.604269 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(1411909703f516361a43617f788315863cba8f0ff9593367629772219adf5dbc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.604300 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(1411909703f516361a43617f788315863cba8f0ff9593367629772219adf5dbc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.604359 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators(3d09cb3e-2427-4282-8112-806df77eb803)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators(3d09cb3e-2427-4282-8112-806df77eb803)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(1411909703f516361a43617f788315863cba8f0ff9593367629772219adf5dbc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" podUID="3d09cb3e-2427-4282-8112-806df77eb803" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.622139 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(481eaf3ba1af1a56dc95fa847b89151e0f15c8cdb844b4e6f37b28300ba4cd9e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.622209 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(481eaf3ba1af1a56dc95fa847b89151e0f15c8cdb844b4e6f37b28300ba4cd9e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.622230 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(481eaf3ba1af1a56dc95fa847b89151e0f15c8cdb844b4e6f37b28300ba4cd9e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.622274 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-67pkr_openshift-operators(91e1df09-581a-4dc8-9dae-6834c2bb3bfe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-67pkr_openshift-operators(91e1df09-581a-4dc8-9dae-6834c2bb3bfe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(481eaf3ba1af1a56dc95fa847b89151e0f15c8cdb844b4e6f37b28300ba4cd9e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-67pkr" podUID="91e1df09-581a-4dc8-9dae-6834c2bb3bfe" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.624527 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(5affab2ab45d0b4ab34f1fae0b9e70528ac96c48ad0dbd3401bfb6cf1053117f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.624597 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(5affab2ab45d0b4ab34f1fae0b9e70528ac96c48ad0dbd3401bfb6cf1053117f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.624620 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(5affab2ab45d0b4ab34f1fae0b9e70528ac96c48ad0dbd3401bfb6cf1053117f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:26 crc kubenswrapper[4894]: E1210 19:07:26.624662 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-89qkl_openshift-operators(641b65b6-0781-479e-ac89-c1cdeb2e5b2d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-89qkl_openshift-operators(641b65b6-0781-479e-ac89-c1cdeb2e5b2d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(5affab2ab45d0b4ab34f1fae0b9e70528ac96c48ad0dbd3401bfb6cf1053117f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" podUID="641b65b6-0781-479e-ac89-c1cdeb2e5b2d" Dec 10 19:07:31 crc kubenswrapper[4894]: I1210 19:07:31.540951 4894 scope.go:117] "RemoveContainer" containerID="d422371c764f725f4e201df30c379baf3507de97186b2eeae087e857e84a3523" Dec 10 19:07:31 crc kubenswrapper[4894]: E1210 19:07:31.541505 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-msfl4_openshift-multus(d9392e0e-ca87-437a-8ba1-9c5d34a50660)\"" pod="openshift-multus/multus-msfl4" podUID="d9392e0e-ca87-437a-8ba1-9c5d34a50660" Dec 10 19:07:37 crc kubenswrapper[4894]: I1210 19:07:37.540731 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:37 crc kubenswrapper[4894]: I1210 19:07:37.543547 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:37 crc kubenswrapper[4894]: E1210 19:07:37.566739 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(2b369a1eb337f8121d91cf1559b83033dc8a4ac78db8c2086e3b40468887f6e5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:37 crc kubenswrapper[4894]: E1210 19:07:37.566818 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(2b369a1eb337f8121d91cf1559b83033dc8a4ac78db8c2086e3b40468887f6e5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:37 crc kubenswrapper[4894]: E1210 19:07:37.566862 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(2b369a1eb337f8121d91cf1559b83033dc8a4ac78db8c2086e3b40468887f6e5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:37 crc kubenswrapper[4894]: E1210 19:07:37.566922 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators(f23b2be9-d333-433e-8381-baec276a1b30)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators(f23b2be9-d333-433e-8381-baec276a1b30)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_openshift-operators_f23b2be9-d333-433e-8381-baec276a1b30_0(2b369a1eb337f8121d91cf1559b83033dc8a4ac78db8c2086e3b40468887f6e5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" podUID="f23b2be9-d333-433e-8381-baec276a1b30" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.540662 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.540721 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.540719 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.540824 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.541313 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.541397 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.541409 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:39 crc kubenswrapper[4894]: I1210 19:07:39.541466 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.621337 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(c16698792b19353c6d500148ec03feab7c30d0906afcc81086b22269a684c75c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.621414 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(c16698792b19353c6d500148ec03feab7c30d0906afcc81086b22269a684c75c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.621442 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(c16698792b19353c6d500148ec03feab7c30d0906afcc81086b22269a684c75c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.621493 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-67pkr_openshift-operators(91e1df09-581a-4dc8-9dae-6834c2bb3bfe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-67pkr_openshift-operators(91e1df09-581a-4dc8-9dae-6834c2bb3bfe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-67pkr_openshift-operators_91e1df09-581a-4dc8-9dae-6834c2bb3bfe_0(c16698792b19353c6d500148ec03feab7c30d0906afcc81086b22269a684c75c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-67pkr" podUID="91e1df09-581a-4dc8-9dae-6834c2bb3bfe" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.628066 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(f9151858cd51894cc1d544a169648cf8b3700649ffca99d0314e037c01af123b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.628135 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(f9151858cd51894cc1d544a169648cf8b3700649ffca99d0314e037c01af123b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.628161 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(f9151858cd51894cc1d544a169648cf8b3700649ffca99d0314e037c01af123b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.628218 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators(3d09cb3e-2427-4282-8112-806df77eb803)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators(3d09cb3e-2427-4282-8112-806df77eb803)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-z6v6g_openshift-operators_3d09cb3e-2427-4282-8112-806df77eb803_0(f9151858cd51894cc1d544a169648cf8b3700649ffca99d0314e037c01af123b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" podUID="3d09cb3e-2427-4282-8112-806df77eb803" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.649048 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1758592c577a37481b59ff69fe69ee8d81515acc5efa454753fa37a972c4a1f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.649143 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1758592c577a37481b59ff69fe69ee8d81515acc5efa454753fa37a972c4a1f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.649169 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1758592c577a37481b59ff69fe69ee8d81515acc5efa454753fa37a972c4a1f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.649227 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-89qkl_openshift-operators(641b65b6-0781-479e-ac89-c1cdeb2e5b2d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-89qkl_openshift-operators(641b65b6-0781-479e-ac89-c1cdeb2e5b2d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-89qkl_openshift-operators_641b65b6-0781-479e-ac89-c1cdeb2e5b2d_0(1758592c577a37481b59ff69fe69ee8d81515acc5efa454753fa37a972c4a1f6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" podUID="641b65b6-0781-479e-ac89-c1cdeb2e5b2d" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.651007 4894 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(a663001b897b4b66af1e454d98004fd8a4a6887b4eede3f92191163637d53e1f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.651091 4894 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(a663001b897b4b66af1e454d98004fd8a4a6887b4eede3f92191163637d53e1f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.651118 4894 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(a663001b897b4b66af1e454d98004fd8a4a6887b4eede3f92191163637d53e1f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:39 crc kubenswrapper[4894]: E1210 19:07:39.651181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators(1b35687b-c053-4902-8ae4-146c82ae486c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators(1b35687b-c053-4902-8ae4-146c82ae486c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_openshift-operators_1b35687b-c053-4902-8ae4-146c82ae486c_0(a663001b897b4b66af1e454d98004fd8a4a6887b4eede3f92191163637d53e1f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" podUID="1b35687b-c053-4902-8ae4-146c82ae486c" Dec 10 19:07:43 crc kubenswrapper[4894]: I1210 19:07:43.542331 4894 scope.go:117] "RemoveContainer" containerID="d422371c764f725f4e201df30c379baf3507de97186b2eeae087e857e84a3523" Dec 10 19:07:44 crc kubenswrapper[4894]: I1210 19:07:44.684776 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-msfl4_d9392e0e-ca87-437a-8ba1-9c5d34a50660/kube-multus/2.log" Dec 10 19:07:44 crc kubenswrapper[4894]: I1210 19:07:44.685203 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-msfl4" event={"ID":"d9392e0e-ca87-437a-8ba1-9c5d34a50660","Type":"ContainerStarted","Data":"9127d19e23398e7978ff4f7759ab52e02641c7ca8e367597431a3eabe334f954"} Dec 10 19:07:47 crc kubenswrapper[4894]: I1210 19:07:47.986093 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rc82k" Dec 10 19:07:48 crc kubenswrapper[4894]: I1210 19:07:48.539833 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:48 crc kubenswrapper[4894]: I1210 19:07:48.540192 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" Dec 10 19:07:48 crc kubenswrapper[4894]: I1210 19:07:48.730708 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj"] Dec 10 19:07:48 crc kubenswrapper[4894]: W1210 19:07:48.742078 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf23b2be9_d333_433e_8381_baec276a1b30.slice/crio-54137ae63003715595d1050b473004fec607a6ab44edafdb055dba08cc70a392 WatchSource:0}: Error finding container 54137ae63003715595d1050b473004fec607a6ab44edafdb055dba08cc70a392: Status 404 returned error can't find the container with id 54137ae63003715595d1050b473004fec607a6ab44edafdb055dba08cc70a392 Dec 10 19:07:49 crc kubenswrapper[4894]: I1210 19:07:49.577451 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:07:49 crc kubenswrapper[4894]: I1210 19:07:49.577747 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:07:49 crc kubenswrapper[4894]: I1210 19:07:49.712251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" event={"ID":"f23b2be9-d333-433e-8381-baec276a1b30","Type":"ContainerStarted","Data":"54137ae63003715595d1050b473004fec607a6ab44edafdb055dba08cc70a392"} Dec 10 19:07:52 crc kubenswrapper[4894]: I1210 19:07:52.540309 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:52 crc kubenswrapper[4894]: I1210 19:07:52.540351 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:52 crc kubenswrapper[4894]: I1210 19:07:52.540309 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:52 crc kubenswrapper[4894]: I1210 19:07:52.540703 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:07:52 crc kubenswrapper[4894]: I1210 19:07:52.540823 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:07:52 crc kubenswrapper[4894]: I1210 19:07:52.541071 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" Dec 10 19:07:55 crc kubenswrapper[4894]: I1210 19:07:55.540179 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:55 crc kubenswrapper[4894]: I1210 19:07:55.541500 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.126117 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-89qkl"] Dec 10 19:07:56 crc kubenswrapper[4894]: W1210 19:07:56.136318 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod641b65b6_0781_479e_ac89_c1cdeb2e5b2d.slice/crio-347577bbc753cc1d7965ad1f9640cc4a67356c6be6a317d8147ae42fb9c1dde7 WatchSource:0}: Error finding container 347577bbc753cc1d7965ad1f9640cc4a67356c6be6a317d8147ae42fb9c1dde7: Status 404 returned error can't find the container with id 347577bbc753cc1d7965ad1f9640cc4a67356c6be6a317d8147ae42fb9c1dde7 Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.422311 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-67pkr"] Dec 10 19:07:56 crc kubenswrapper[4894]: W1210 19:07:56.427465 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91e1df09_581a_4dc8_9dae_6834c2bb3bfe.slice/crio-583aec9958bc1730ddee006c32c0ff10e42e84dac12ff10300ab054eea6af261 WatchSource:0}: Error finding container 583aec9958bc1730ddee006c32c0ff10e42e84dac12ff10300ab054eea6af261: Status 404 returned error can't find the container with id 583aec9958bc1730ddee006c32c0ff10e42e84dac12ff10300ab054eea6af261 Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.476577 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk"] Dec 10 19:07:56 crc kubenswrapper[4894]: W1210 19:07:56.477004 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b35687b_c053_4902_8ae4_146c82ae486c.slice/crio-d0bd351d4d63b268478d1dc54a6099afaff4edf6740d01f8b53d8e8554cbf0ce WatchSource:0}: Error finding container d0bd351d4d63b268478d1dc54a6099afaff4edf6740d01f8b53d8e8554cbf0ce: Status 404 returned error can't find the container with id d0bd351d4d63b268478d1dc54a6099afaff4edf6740d01f8b53d8e8554cbf0ce Dec 10 19:07:56 crc kubenswrapper[4894]: W1210 19:07:56.477956 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d09cb3e_2427_4282_8112_806df77eb803.slice/crio-a1bcb8120d04ab18c00e31b61c4d581f79aa3a019209da0bb56aa29f5725c9a2 WatchSource:0}: Error finding container a1bcb8120d04ab18c00e31b61c4d581f79aa3a019209da0bb56aa29f5725c9a2: Status 404 returned error can't find the container with id a1bcb8120d04ab18c00e31b61c4d581f79aa3a019209da0bb56aa29f5725c9a2 Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.480882 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g"] Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.756189 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" event={"ID":"3d09cb3e-2427-4282-8112-806df77eb803","Type":"ContainerStarted","Data":"a1bcb8120d04ab18c00e31b61c4d581f79aa3a019209da0bb56aa29f5725c9a2"} Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.757842 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" event={"ID":"1b35687b-c053-4902-8ae4-146c82ae486c","Type":"ContainerStarted","Data":"da5bf6260867c81ac23fe2f62c3e0b8a637e0c2334d8076478ded2ca4ab2ed11"} Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.757919 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" event={"ID":"1b35687b-c053-4902-8ae4-146c82ae486c","Type":"ContainerStarted","Data":"d0bd351d4d63b268478d1dc54a6099afaff4edf6740d01f8b53d8e8554cbf0ce"} Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.759579 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-67pkr" event={"ID":"91e1df09-581a-4dc8-9dae-6834c2bb3bfe","Type":"ContainerStarted","Data":"583aec9958bc1730ddee006c32c0ff10e42e84dac12ff10300ab054eea6af261"} Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.760585 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" event={"ID":"641b65b6-0781-479e-ac89-c1cdeb2e5b2d","Type":"ContainerStarted","Data":"347577bbc753cc1d7965ad1f9640cc4a67356c6be6a317d8147ae42fb9c1dde7"} Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.762143 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" event={"ID":"f23b2be9-d333-433e-8381-baec276a1b30","Type":"ContainerStarted","Data":"426741daa3208bbfcc04c1c563f93cfa1488c49b17a9c99fe53c58f5a1b77397"} Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.779407 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk" podStartSLOduration=33.779382332 podStartE2EDuration="33.779382332s" podCreationTimestamp="2025-12-10 19:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:07:56.774654996 +0000 UTC m=+759.569502798" watchObservedRunningTime="2025-12-10 19:07:56.779382332 +0000 UTC m=+759.574230144" Dec 10 19:07:56 crc kubenswrapper[4894]: I1210 19:07:56.845807 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj" podStartSLOduration=26.576461114 podStartE2EDuration="33.845783397s" podCreationTimestamp="2025-12-10 19:07:23 +0000 UTC" firstStartedPulling="2025-12-10 19:07:48.74421707 +0000 UTC m=+751.539064872" lastFinishedPulling="2025-12-10 19:07:56.013539363 +0000 UTC m=+758.808387155" observedRunningTime="2025-12-10 19:07:56.842301354 +0000 UTC m=+759.637149236" watchObservedRunningTime="2025-12-10 19:07:56.845783397 +0000 UTC m=+759.640631229" Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.814296 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-67pkr" event={"ID":"91e1df09-581a-4dc8-9dae-6834c2bb3bfe","Type":"ContainerStarted","Data":"7ce22367a5023ebe44d5dc9394f563052862b4d235cd85bf9efed6330cbfeef2"} Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.815066 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.817005 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" event={"ID":"641b65b6-0781-479e-ac89-c1cdeb2e5b2d","Type":"ContainerStarted","Data":"37771577d9e235c90fab44ee8e7eb33516f61a830eb8178279cc9841a984c4aa"} Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.817199 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.818867 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" event={"ID":"3d09cb3e-2427-4282-8112-806df77eb803","Type":"ContainerStarted","Data":"2b5cadd1e4b3f48f4538be6336b8ed9b6d1fc1d967d26f7670bb463bcaca2f53"} Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.821299 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.838538 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-67pkr" podStartSLOduration=33.95548355 podStartE2EDuration="39.838517783s" podCreationTimestamp="2025-12-10 19:07:23 +0000 UTC" firstStartedPulling="2025-12-10 19:07:56.4307825 +0000 UTC m=+759.225630292" lastFinishedPulling="2025-12-10 19:08:02.313816733 +0000 UTC m=+765.108664525" observedRunningTime="2025-12-10 19:08:02.835245516 +0000 UTC m=+765.630093328" watchObservedRunningTime="2025-12-10 19:08:02.838517783 +0000 UTC m=+765.633365595" Dec 10 19:08:02 crc kubenswrapper[4894]: I1210 19:08:02.858459 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-z6v6g" podStartSLOduration=34.027780635 podStartE2EDuration="39.858442447s" podCreationTimestamp="2025-12-10 19:07:23 +0000 UTC" firstStartedPulling="2025-12-10 19:07:56.480049627 +0000 UTC m=+759.274897419" lastFinishedPulling="2025-12-10 19:08:02.310711439 +0000 UTC m=+765.105559231" observedRunningTime="2025-12-10 19:08:02.855924299 +0000 UTC m=+765.650772121" watchObservedRunningTime="2025-12-10 19:08:02.858442447 +0000 UTC m=+765.653290239" Dec 10 19:08:06 crc kubenswrapper[4894]: I1210 19:08:06.548330 4894 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.848712 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-89qkl" podStartSLOduration=39.650118236 podStartE2EDuration="45.848691866s" podCreationTimestamp="2025-12-10 19:07:23 +0000 UTC" firstStartedPulling="2025-12-10 19:07:56.138759291 +0000 UTC m=+758.933607083" lastFinishedPulling="2025-12-10 19:08:02.337332921 +0000 UTC m=+765.132180713" observedRunningTime="2025-12-10 19:08:02.893828653 +0000 UTC m=+765.688676465" watchObservedRunningTime="2025-12-10 19:08:08.848691866 +0000 UTC m=+771.643539678" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.851482 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qv9lh"] Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.852368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.855646 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vqs2f"] Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.856590 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vqs2f" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.862604 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.862867 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-ksns8" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.862955 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.866742 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-s64vc" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.882940 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vqs2f"] Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.889872 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qv9lh"] Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.933885 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fsvcg"] Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.958202 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.966223 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fsvcg"] Dec 10 19:08:08 crc kubenswrapper[4894]: I1210 19:08:08.982167 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-bd99p" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.013982 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2qv9\" (UniqueName: \"kubernetes.io/projected/451bafa9-45d0-4374-8bad-29067fd16d26-kube-api-access-h2qv9\") pod \"cert-manager-5b446d88c5-vqs2f\" (UID: \"451bafa9-45d0-4374-8bad-29067fd16d26\") " pod="cert-manager/cert-manager-5b446d88c5-vqs2f" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.014290 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2cc6\" (UniqueName: \"kubernetes.io/projected/1134d9d0-0ca0-4a51-8066-acc409fd3218-kube-api-access-q2cc6\") pod \"cert-manager-cainjector-7f985d654d-qv9lh\" (UID: \"1134d9d0-0ca0-4a51-8066-acc409fd3218\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.115549 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2cc6\" (UniqueName: \"kubernetes.io/projected/1134d9d0-0ca0-4a51-8066-acc409fd3218-kube-api-access-q2cc6\") pod \"cert-manager-cainjector-7f985d654d-qv9lh\" (UID: \"1134d9d0-0ca0-4a51-8066-acc409fd3218\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.115621 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc6l8\" (UniqueName: \"kubernetes.io/projected/ca50f7b6-2bd0-48ce-af0c-9edf0b895058-kube-api-access-gc6l8\") pod \"cert-manager-webhook-5655c58dd6-fsvcg\" (UID: \"ca50f7b6-2bd0-48ce-af0c-9edf0b895058\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.115660 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2qv9\" (UniqueName: \"kubernetes.io/projected/451bafa9-45d0-4374-8bad-29067fd16d26-kube-api-access-h2qv9\") pod \"cert-manager-5b446d88c5-vqs2f\" (UID: \"451bafa9-45d0-4374-8bad-29067fd16d26\") " pod="cert-manager/cert-manager-5b446d88c5-vqs2f" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.135221 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2cc6\" (UniqueName: \"kubernetes.io/projected/1134d9d0-0ca0-4a51-8066-acc409fd3218-kube-api-access-q2cc6\") pod \"cert-manager-cainjector-7f985d654d-qv9lh\" (UID: \"1134d9d0-0ca0-4a51-8066-acc409fd3218\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.136998 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2qv9\" (UniqueName: \"kubernetes.io/projected/451bafa9-45d0-4374-8bad-29067fd16d26-kube-api-access-h2qv9\") pod \"cert-manager-5b446d88c5-vqs2f\" (UID: \"451bafa9-45d0-4374-8bad-29067fd16d26\") " pod="cert-manager/cert-manager-5b446d88c5-vqs2f" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.175088 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.180370 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vqs2f" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.217217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc6l8\" (UniqueName: \"kubernetes.io/projected/ca50f7b6-2bd0-48ce-af0c-9edf0b895058-kube-api-access-gc6l8\") pod \"cert-manager-webhook-5655c58dd6-fsvcg\" (UID: \"ca50f7b6-2bd0-48ce-af0c-9edf0b895058\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.236446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc6l8\" (UniqueName: \"kubernetes.io/projected/ca50f7b6-2bd0-48ce-af0c-9edf0b895058-kube-api-access-gc6l8\") pod \"cert-manager-webhook-5655c58dd6-fsvcg\" (UID: \"ca50f7b6-2bd0-48ce-af0c-9edf0b895058\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.278333 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.397304 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vqs2f"] Dec 10 19:08:09 crc kubenswrapper[4894]: W1210 19:08:09.404300 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod451bafa9_45d0_4374_8bad_29067fd16d26.slice/crio-aea2d85720dded121f4d32045f145f6439840129880ea7f89bcf3535523d0a96 WatchSource:0}: Error finding container aea2d85720dded121f4d32045f145f6439840129880ea7f89bcf3535523d0a96: Status 404 returned error can't find the container with id aea2d85720dded121f4d32045f145f6439840129880ea7f89bcf3535523d0a96 Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.426958 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qv9lh"] Dec 10 19:08:09 crc kubenswrapper[4894]: W1210 19:08:09.433974 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1134d9d0_0ca0_4a51_8066_acc409fd3218.slice/crio-38dfa45d81fd37b99f9bd4acb3e87b077f077ac81bb38c19f6185904c27f19e4 WatchSource:0}: Error finding container 38dfa45d81fd37b99f9bd4acb3e87b077f077ac81bb38c19f6185904c27f19e4: Status 404 returned error can't find the container with id 38dfa45d81fd37b99f9bd4acb3e87b077f077ac81bb38c19f6185904c27f19e4 Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.479261 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fsvcg"] Dec 10 19:08:09 crc kubenswrapper[4894]: W1210 19:08:09.482117 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca50f7b6_2bd0_48ce_af0c_9edf0b895058.slice/crio-14157331ff6c87fdb175f5d54cf84c0708ab12946e90d21751f3b959aa35ccbd WatchSource:0}: Error finding container 14157331ff6c87fdb175f5d54cf84c0708ab12946e90d21751f3b959aa35ccbd: Status 404 returned error can't find the container with id 14157331ff6c87fdb175f5d54cf84c0708ab12946e90d21751f3b959aa35ccbd Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.865301 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" event={"ID":"ca50f7b6-2bd0-48ce-af0c-9edf0b895058","Type":"ContainerStarted","Data":"14157331ff6c87fdb175f5d54cf84c0708ab12946e90d21751f3b959aa35ccbd"} Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.867107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vqs2f" event={"ID":"451bafa9-45d0-4374-8bad-29067fd16d26","Type":"ContainerStarted","Data":"aea2d85720dded121f4d32045f145f6439840129880ea7f89bcf3535523d0a96"} Dec 10 19:08:09 crc kubenswrapper[4894]: I1210 19:08:09.869445 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" event={"ID":"1134d9d0-0ca0-4a51-8066-acc409fd3218","Type":"ContainerStarted","Data":"38dfa45d81fd37b99f9bd4acb3e87b077f077ac81bb38c19f6185904c27f19e4"} Dec 10 19:08:11 crc kubenswrapper[4894]: I1210 19:08:11.881598 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vqs2f" event={"ID":"451bafa9-45d0-4374-8bad-29067fd16d26","Type":"ContainerStarted","Data":"8f51fbc46184ec041d28b44562a71abe3a5dbafdd906d19e2555bdcc5a9c865b"} Dec 10 19:08:11 crc kubenswrapper[4894]: I1210 19:08:11.902691 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vqs2f" podStartSLOduration=1.887998548 podStartE2EDuration="3.90267672s" podCreationTimestamp="2025-12-10 19:08:08 +0000 UTC" firstStartedPulling="2025-12-10 19:08:09.406535523 +0000 UTC m=+772.201383305" lastFinishedPulling="2025-12-10 19:08:11.421213675 +0000 UTC m=+774.216061477" observedRunningTime="2025-12-10 19:08:11.902094274 +0000 UTC m=+774.696942076" watchObservedRunningTime="2025-12-10 19:08:11.90267672 +0000 UTC m=+774.697524512" Dec 10 19:08:13 crc kubenswrapper[4894]: I1210 19:08:13.909791 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" event={"ID":"1134d9d0-0ca0-4a51-8066-acc409fd3218","Type":"ContainerStarted","Data":"1c4e89f79e5fb79a3150b9369b0ffe46cf39f561a4e5127be1536faa564352f5"} Dec 10 19:08:13 crc kubenswrapper[4894]: I1210 19:08:13.912674 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" event={"ID":"ca50f7b6-2bd0-48ce-af0c-9edf0b895058","Type":"ContainerStarted","Data":"127f8d761e7edbdeaa363fee759211f16472c621a61ac26def16679ff8d10d87"} Dec 10 19:08:13 crc kubenswrapper[4894]: I1210 19:08:13.913460 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:13 crc kubenswrapper[4894]: I1210 19:08:13.944444 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qv9lh" podStartSLOduration=2.5053651649999997 podStartE2EDuration="5.944423556s" podCreationTimestamp="2025-12-10 19:08:08 +0000 UTC" firstStartedPulling="2025-12-10 19:08:09.439961846 +0000 UTC m=+772.234809628" lastFinishedPulling="2025-12-10 19:08:12.879020197 +0000 UTC m=+775.673868019" observedRunningTime="2025-12-10 19:08:13.939511445 +0000 UTC m=+776.734359277" watchObservedRunningTime="2025-12-10 19:08:13.944423556 +0000 UTC m=+776.739271368" Dec 10 19:08:13 crc kubenswrapper[4894]: I1210 19:08:13.956471 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" podStartSLOduration=2.562985566 podStartE2EDuration="5.956449678s" podCreationTimestamp="2025-12-10 19:08:08 +0000 UTC" firstStartedPulling="2025-12-10 19:08:09.48421853 +0000 UTC m=+772.279066322" lastFinishedPulling="2025-12-10 19:08:12.877682642 +0000 UTC m=+775.672530434" observedRunningTime="2025-12-10 19:08:13.954059244 +0000 UTC m=+776.748907036" watchObservedRunningTime="2025-12-10 19:08:13.956449678 +0000 UTC m=+776.751297480" Dec 10 19:08:15 crc kubenswrapper[4894]: I1210 19:08:15.800508 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-67pkr" Dec 10 19:08:19 crc kubenswrapper[4894]: I1210 19:08:19.280893 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsvcg" Dec 10 19:08:19 crc kubenswrapper[4894]: I1210 19:08:19.576974 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:08:19 crc kubenswrapper[4894]: I1210 19:08:19.577021 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.611751 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9"] Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.613477 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.620212 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.643618 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9"] Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.736023 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jhnj\" (UniqueName: \"kubernetes.io/projected/6fa7084b-6307-4bbb-b690-076cc71fb40c-kube-api-access-8jhnj\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.736159 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-bundle\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.736239 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-util\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.837492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-util\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.837559 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jhnj\" (UniqueName: \"kubernetes.io/projected/6fa7084b-6307-4bbb-b690-076cc71fb40c-kube-api-access-8jhnj\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.837605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-bundle\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.838068 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-bundle\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.838287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-util\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.872052 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jhnj\" (UniqueName: \"kubernetes.io/projected/6fa7084b-6307-4bbb-b690-076cc71fb40c-kube-api-access-8jhnj\") pod \"3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:45 crc kubenswrapper[4894]: I1210 19:08:45.931830 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:46 crc kubenswrapper[4894]: I1210 19:08:46.380159 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9"] Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.160868 4894 generic.go:334] "Generic (PLEG): container finished" podID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerID="e48c386789ddad15e3a27d1db1cc58c923d9ea4a0f6807fce2e7f642a73292ed" exitCode=0 Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.160936 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" event={"ID":"6fa7084b-6307-4bbb-b690-076cc71fb40c","Type":"ContainerDied","Data":"e48c386789ddad15e3a27d1db1cc58c923d9ea4a0f6807fce2e7f642a73292ed"} Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.161190 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" event={"ID":"6fa7084b-6307-4bbb-b690-076cc71fb40c","Type":"ContainerStarted","Data":"c60eaafb79774b1a0f6e4f70760dff46405a692475fb77852762210fe204bd90"} Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.723510 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ld7sc"] Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.724872 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.740822 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ld7sc"] Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.868124 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-catalog-content\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.868189 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfxx\" (UniqueName: \"kubernetes.io/projected/5da3760f-e20c-4218-9ed0-1af29ec2c41f-kube-api-access-lzfxx\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.868228 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-utilities\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.969140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-catalog-content\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.969194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfxx\" (UniqueName: \"kubernetes.io/projected/5da3760f-e20c-4218-9ed0-1af29ec2c41f-kube-api-access-lzfxx\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.969230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-utilities\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.969701 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-utilities\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:47 crc kubenswrapper[4894]: I1210 19:08:47.969815 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-catalog-content\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.001664 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfxx\" (UniqueName: \"kubernetes.io/projected/5da3760f-e20c-4218-9ed0-1af29ec2c41f-kube-api-access-lzfxx\") pod \"redhat-operators-ld7sc\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.047521 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.048989 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.051450 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.051685 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.051877 4894 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-dkmjf" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.062480 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.077511 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.171135 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") " pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.171168 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5b88\" (UniqueName: \"kubernetes.io/projected/592cdea9-b78f-4d9c-b932-e3960d806640-kube-api-access-k5b88\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") " pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.271819 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") " pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.271901 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5b88\" (UniqueName: \"kubernetes.io/projected/592cdea9-b78f-4d9c-b932-e3960d806640-kube-api-access-k5b88\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") " pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.275353 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.275405 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/65eb30749b818dc0864d7549ff3b09f9c71ea1e2978b7c1b20491c4f5cce0faa/globalmount\"" pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.299119 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5b88\" (UniqueName: \"kubernetes.io/projected/592cdea9-b78f-4d9c-b932-e3960d806640-kube-api-access-k5b88\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") " pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.305526 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ld7sc"] Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.343113 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3f3b785c-7e3c-4c91-9150-4ccae022a05b\") pod \"minio\" (UID: \"592cdea9-b78f-4d9c-b932-e3960d806640\") " pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.363120 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 10 19:08:48 crc kubenswrapper[4894]: I1210 19:08:48.586062 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 10 19:08:48 crc kubenswrapper[4894]: W1210 19:08:48.602659 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod592cdea9_b78f_4d9c_b932_e3960d806640.slice/crio-6089b0e08bea98200c1f29a526c3b1c58d464dd4c4966da5c712dbeaae659239 WatchSource:0}: Error finding container 6089b0e08bea98200c1f29a526c3b1c58d464dd4c4966da5c712dbeaae659239: Status 404 returned error can't find the container with id 6089b0e08bea98200c1f29a526c3b1c58d464dd4c4966da5c712dbeaae659239 Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.174760 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"592cdea9-b78f-4d9c-b932-e3960d806640","Type":"ContainerStarted","Data":"6089b0e08bea98200c1f29a526c3b1c58d464dd4c4966da5c712dbeaae659239"} Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.176510 4894 generic.go:334] "Generic (PLEG): container finished" podID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerID="74b9fe3afadf17eccd77366e8d5c25c6a9a25d5e570feb4b0843d47c930bacdd" exitCode=0 Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.176593 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" event={"ID":"6fa7084b-6307-4bbb-b690-076cc71fb40c","Type":"ContainerDied","Data":"74b9fe3afadf17eccd77366e8d5c25c6a9a25d5e570feb4b0843d47c930bacdd"} Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.178628 4894 generic.go:334] "Generic (PLEG): container finished" podID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerID="3afe1a3074eaa049b0248b64fb40b14814f02ddfd545f5138ddedde0c9a61e7c" exitCode=0 Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.178659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerDied","Data":"3afe1a3074eaa049b0248b64fb40b14814f02ddfd545f5138ddedde0c9a61e7c"} Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.178681 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerStarted","Data":"e5c080d4dadbfee6c2c6f908b5f783df867c53ebce9b17a6490146b9962032d6"} Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.576801 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.577216 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.577261 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.578263 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ec2353aa9d8d229a346f66ddc339bc5b86fce1f5b3bf7d51741a93c0c96e096"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:08:49 crc kubenswrapper[4894]: I1210 19:08:49.578317 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://4ec2353aa9d8d229a346f66ddc339bc5b86fce1f5b3bf7d51741a93c0c96e096" gracePeriod=600 Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.188110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerStarted","Data":"a2e9e39fb8f79c93ca5a57ebfd996f3017d8c0d8d6f980d7a59136f2244e9b76"} Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.190239 4894 generic.go:334] "Generic (PLEG): container finished" podID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerID="ac8ee36044dea5ff0b71d35033a9001a64e879f796efdf1ad0a42ebf27ba243e" exitCode=0 Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.190315 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" event={"ID":"6fa7084b-6307-4bbb-b690-076cc71fb40c","Type":"ContainerDied","Data":"ac8ee36044dea5ff0b71d35033a9001a64e879f796efdf1ad0a42ebf27ba243e"} Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.193146 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="4ec2353aa9d8d229a346f66ddc339bc5b86fce1f5b3bf7d51741a93c0c96e096" exitCode=0 Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.193187 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"4ec2353aa9d8d229a346f66ddc339bc5b86fce1f5b3bf7d51741a93c0c96e096"} Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.193208 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"a0a333f57d00f6ace0bfa8b35a6220c82d5630155e1eac3a22921478ab8cbce3"} Dec 10 19:08:50 crc kubenswrapper[4894]: I1210 19:08:50.193224 4894 scope.go:117] "RemoveContainer" containerID="1143033875ce2861ff6783d3b0a8fe27cde229985e69e12e38acac4bb01e1718" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.290001 4894 generic.go:334] "Generic (PLEG): container finished" podID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerID="a2e9e39fb8f79c93ca5a57ebfd996f3017d8c0d8d6f980d7a59136f2244e9b76" exitCode=0 Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.290154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerDied","Data":"a2e9e39fb8f79c93ca5a57ebfd996f3017d8c0d8d6f980d7a59136f2244e9b76"} Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.744200 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.787221 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jhnj\" (UniqueName: \"kubernetes.io/projected/6fa7084b-6307-4bbb-b690-076cc71fb40c-kube-api-access-8jhnj\") pod \"6fa7084b-6307-4bbb-b690-076cc71fb40c\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.787345 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-util\") pod \"6fa7084b-6307-4bbb-b690-076cc71fb40c\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.787471 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-bundle\") pod \"6fa7084b-6307-4bbb-b690-076cc71fb40c\" (UID: \"6fa7084b-6307-4bbb-b690-076cc71fb40c\") " Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.794074 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-bundle" (OuterVolumeSpecName: "bundle") pod "6fa7084b-6307-4bbb-b690-076cc71fb40c" (UID: "6fa7084b-6307-4bbb-b690-076cc71fb40c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.808268 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-util" (OuterVolumeSpecName: "util") pod "6fa7084b-6307-4bbb-b690-076cc71fb40c" (UID: "6fa7084b-6307-4bbb-b690-076cc71fb40c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.817059 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa7084b-6307-4bbb-b690-076cc71fb40c-kube-api-access-8jhnj" (OuterVolumeSpecName: "kube-api-access-8jhnj") pod "6fa7084b-6307-4bbb-b690-076cc71fb40c" (UID: "6fa7084b-6307-4bbb-b690-076cc71fb40c"). InnerVolumeSpecName "kube-api-access-8jhnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.889046 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.889086 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jhnj\" (UniqueName: \"kubernetes.io/projected/6fa7084b-6307-4bbb-b690-076cc71fb40c-kube-api-access-8jhnj\") on node \"crc\" DevicePath \"\"" Dec 10 19:08:52 crc kubenswrapper[4894]: I1210 19:08:52.889101 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6fa7084b-6307-4bbb-b690-076cc71fb40c-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:08:53 crc kubenswrapper[4894]: I1210 19:08:53.299862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" event={"ID":"6fa7084b-6307-4bbb-b690-076cc71fb40c","Type":"ContainerDied","Data":"c60eaafb79774b1a0f6e4f70760dff46405a692475fb77852762210fe204bd90"} Dec 10 19:08:53 crc kubenswrapper[4894]: I1210 19:08:53.299899 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c60eaafb79774b1a0f6e4f70760dff46405a692475fb77852762210fe204bd90" Dec 10 19:08:53 crc kubenswrapper[4894]: I1210 19:08:53.299943 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9" Dec 10 19:08:54 crc kubenswrapper[4894]: I1210 19:08:54.307902 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerStarted","Data":"c062f6116489b38046d532e7b8173c854ee2a79160b6c83301df0ad1ec043027"} Dec 10 19:08:54 crc kubenswrapper[4894]: I1210 19:08:54.309671 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"592cdea9-b78f-4d9c-b932-e3960d806640","Type":"ContainerStarted","Data":"01cfca8809e9cf8cde9c83839138c77d9f924ef9ae79c436d12534cd13600110"} Dec 10 19:08:54 crc kubenswrapper[4894]: I1210 19:08:54.326590 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ld7sc" podStartSLOduration=3.055187703 podStartE2EDuration="7.32657205s" podCreationTimestamp="2025-12-10 19:08:47 +0000 UTC" firstStartedPulling="2025-12-10 19:08:49.179903278 +0000 UTC m=+811.974751070" lastFinishedPulling="2025-12-10 19:08:53.451287625 +0000 UTC m=+816.246135417" observedRunningTime="2025-12-10 19:08:54.325922873 +0000 UTC m=+817.120770675" watchObservedRunningTime="2025-12-10 19:08:54.32657205 +0000 UTC m=+817.121419842" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.077883 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.078201 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.576752 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=8.719039935 podStartE2EDuration="13.57673234s" podCreationTimestamp="2025-12-10 19:08:45 +0000 UTC" firstStartedPulling="2025-12-10 19:08:48.60555885 +0000 UTC m=+811.400406642" lastFinishedPulling="2025-12-10 19:08:53.463251255 +0000 UTC m=+816.258099047" observedRunningTime="2025-12-10 19:08:54.341227342 +0000 UTC m=+817.136075134" watchObservedRunningTime="2025-12-10 19:08:58.57673234 +0000 UTC m=+821.371580152" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.579920 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk"] Dec 10 19:08:58 crc kubenswrapper[4894]: E1210 19:08:58.580191 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="pull" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.580214 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="pull" Dec 10 19:08:58 crc kubenswrapper[4894]: E1210 19:08:58.580235 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="util" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.580244 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="util" Dec 10 19:08:58 crc kubenswrapper[4894]: E1210 19:08:58.580257 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="extract" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.580264 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="extract" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.580388 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa7084b-6307-4bbb-b690-076cc71fb40c" containerName="extract" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.581110 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.584004 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.584222 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.588936 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-8bgn6" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.589007 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.589217 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.591579 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.661657 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk"] Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.762609 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmtcj\" (UniqueName: \"kubernetes.io/projected/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-kube-api-access-bmtcj\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.762659 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-webhook-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.762708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-apiservice-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.762733 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.762767 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-manager-config\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.864050 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmtcj\" (UniqueName: \"kubernetes.io/projected/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-kube-api-access-bmtcj\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.864102 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-webhook-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.864143 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-apiservice-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.864164 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.864193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-manager-config\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.865385 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-manager-config\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.871557 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.871596 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-webhook-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.873334 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-apiservice-cert\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.881580 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmtcj\" (UniqueName: \"kubernetes.io/projected/c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec-kube-api-access-bmtcj\") pod \"loki-operator-controller-manager-59fd4b8cb8-vkdlk\" (UID: \"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec\") " pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:58 crc kubenswrapper[4894]: I1210 19:08:58.908237 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:08:59 crc kubenswrapper[4894]: I1210 19:08:59.172109 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ld7sc" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="registry-server" probeResult="failure" output=< Dec 10 19:08:59 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:08:59 crc kubenswrapper[4894]: > Dec 10 19:08:59 crc kubenswrapper[4894]: I1210 19:08:59.188725 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk"] Dec 10 19:08:59 crc kubenswrapper[4894]: I1210 19:08:59.337366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" event={"ID":"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec","Type":"ContainerStarted","Data":"25162f5a4a55bb7224871fb8aab148e7ccb8d57be5e453932c04ecd8e6f694a8"} Dec 10 19:09:05 crc kubenswrapper[4894]: I1210 19:09:05.381467 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" event={"ID":"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec","Type":"ContainerStarted","Data":"a6dc70cc2574342c008f0c8048a67004b0e902d437c422a0f23db96f5104753d"} Dec 10 19:09:08 crc kubenswrapper[4894]: I1210 19:09:08.139141 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:09:08 crc kubenswrapper[4894]: I1210 19:09:08.181521 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:09:10 crc kubenswrapper[4894]: I1210 19:09:10.107615 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ld7sc"] Dec 10 19:09:10 crc kubenswrapper[4894]: I1210 19:09:10.108162 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ld7sc" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="registry-server" containerID="cri-o://c062f6116489b38046d532e7b8173c854ee2a79160b6c83301df0ad1ec043027" gracePeriod=2 Dec 10 19:09:10 crc kubenswrapper[4894]: I1210 19:09:10.415080 4894 generic.go:334] "Generic (PLEG): container finished" podID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerID="c062f6116489b38046d532e7b8173c854ee2a79160b6c83301df0ad1ec043027" exitCode=0 Dec 10 19:09:10 crc kubenswrapper[4894]: I1210 19:09:10.415134 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerDied","Data":"c062f6116489b38046d532e7b8173c854ee2a79160b6c83301df0ad1ec043027"} Dec 10 19:09:10 crc kubenswrapper[4894]: I1210 19:09:10.923828 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.065372 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-utilities\") pod \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.065507 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzfxx\" (UniqueName: \"kubernetes.io/projected/5da3760f-e20c-4218-9ed0-1af29ec2c41f-kube-api-access-lzfxx\") pod \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.065560 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-catalog-content\") pod \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\" (UID: \"5da3760f-e20c-4218-9ed0-1af29ec2c41f\") " Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.067832 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-utilities" (OuterVolumeSpecName: "utilities") pod "5da3760f-e20c-4218-9ed0-1af29ec2c41f" (UID: "5da3760f-e20c-4218-9ed0-1af29ec2c41f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.073207 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da3760f-e20c-4218-9ed0-1af29ec2c41f-kube-api-access-lzfxx" (OuterVolumeSpecName: "kube-api-access-lzfxx") pod "5da3760f-e20c-4218-9ed0-1af29ec2c41f" (UID: "5da3760f-e20c-4218-9ed0-1af29ec2c41f"). InnerVolumeSpecName "kube-api-access-lzfxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.167343 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.167389 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzfxx\" (UniqueName: \"kubernetes.io/projected/5da3760f-e20c-4218-9ed0-1af29ec2c41f-kube-api-access-lzfxx\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.202658 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5da3760f-e20c-4218-9ed0-1af29ec2c41f" (UID: "5da3760f-e20c-4218-9ed0-1af29ec2c41f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.268643 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5da3760f-e20c-4218-9ed0-1af29ec2c41f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.424032 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ld7sc" event={"ID":"5da3760f-e20c-4218-9ed0-1af29ec2c41f","Type":"ContainerDied","Data":"e5c080d4dadbfee6c2c6f908b5f783df867c53ebce9b17a6490146b9962032d6"} Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.424083 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ld7sc" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.424129 4894 scope.go:117] "RemoveContainer" containerID="c062f6116489b38046d532e7b8173c854ee2a79160b6c83301df0ad1ec043027" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.426310 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" event={"ID":"c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec","Type":"ContainerStarted","Data":"07bd2a1e3f8bf79c50ed764fcefb819f75671809f24d57e08601ead273d8665e"} Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.426596 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.429576 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.461960 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-59fd4b8cb8-vkdlk" podStartSLOduration=1.878841298 podStartE2EDuration="13.461934671s" podCreationTimestamp="2025-12-10 19:08:58 +0000 UTC" firstStartedPulling="2025-12-10 19:08:59.205535384 +0000 UTC m=+822.000383186" lastFinishedPulling="2025-12-10 19:09:10.788628747 +0000 UTC m=+833.583476559" observedRunningTime="2025-12-10 19:09:11.461251224 +0000 UTC m=+834.256099076" watchObservedRunningTime="2025-12-10 19:09:11.461934671 +0000 UTC m=+834.256782503" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.469448 4894 scope.go:117] "RemoveContainer" containerID="a2e9e39fb8f79c93ca5a57ebfd996f3017d8c0d8d6f980d7a59136f2244e9b76" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.500242 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ld7sc"] Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.517664 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ld7sc"] Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.522005 4894 scope.go:117] "RemoveContainer" containerID="3afe1a3074eaa049b0248b64fb40b14814f02ddfd545f5138ddedde0c9a61e7c" Dec 10 19:09:11 crc kubenswrapper[4894]: I1210 19:09:11.559792 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" path="/var/lib/kubelet/pods/5da3760f-e20c-4218-9ed0-1af29ec2c41f/volumes" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.866578 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4"] Dec 10 19:09:44 crc kubenswrapper[4894]: E1210 19:09:44.867365 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="extract-content" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.867380 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="extract-content" Dec 10 19:09:44 crc kubenswrapper[4894]: E1210 19:09:44.867401 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="extract-utilities" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.867410 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="extract-utilities" Dec 10 19:09:44 crc kubenswrapper[4894]: E1210 19:09:44.867429 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="registry-server" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.867438 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="registry-server" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.867567 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da3760f-e20c-4218-9ed0-1af29ec2c41f" containerName="registry-server" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.868540 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.870923 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.876214 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4"] Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.913910 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.913966 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftnlf\" (UniqueName: \"kubernetes.io/projected/877cd716-19e7-461e-b218-711dafd4f785-kube-api-access-ftnlf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:44 crc kubenswrapper[4894]: I1210 19:09:44.913992 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.014712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.014771 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftnlf\" (UniqueName: \"kubernetes.io/projected/877cd716-19e7-461e-b218-711dafd4f785-kube-api-access-ftnlf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.014800 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.015381 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.015572 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.038735 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftnlf\" (UniqueName: \"kubernetes.io/projected/877cd716-19e7-461e-b218-711dafd4f785-kube-api-access-ftnlf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.182652 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.425906 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4"] Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.678572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" event={"ID":"877cd716-19e7-461e-b218-711dafd4f785","Type":"ContainerStarted","Data":"4993b066044152536c574330b44548616f0e7c48e765e5eb0556242c81e0c014"} Dec 10 19:09:45 crc kubenswrapper[4894]: I1210 19:09:45.678928 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" event={"ID":"877cd716-19e7-461e-b218-711dafd4f785","Type":"ContainerStarted","Data":"529a2c3a155b3499b6f62420267d07261f39313c165689e688e21edd75ce9d1b"} Dec 10 19:09:46 crc kubenswrapper[4894]: I1210 19:09:46.691018 4894 generic.go:334] "Generic (PLEG): container finished" podID="877cd716-19e7-461e-b218-711dafd4f785" containerID="4993b066044152536c574330b44548616f0e7c48e765e5eb0556242c81e0c014" exitCode=0 Dec 10 19:09:46 crc kubenswrapper[4894]: I1210 19:09:46.691100 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" event={"ID":"877cd716-19e7-461e-b218-711dafd4f785","Type":"ContainerDied","Data":"4993b066044152536c574330b44548616f0e7c48e765e5eb0556242c81e0c014"} Dec 10 19:09:48 crc kubenswrapper[4894]: I1210 19:09:48.708572 4894 generic.go:334] "Generic (PLEG): container finished" podID="877cd716-19e7-461e-b218-711dafd4f785" containerID="05af8c0b0757ab5358234087dfa025f5fe7f5339bb238dd5260ff8748c56647d" exitCode=0 Dec 10 19:09:48 crc kubenswrapper[4894]: I1210 19:09:48.708632 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" event={"ID":"877cd716-19e7-461e-b218-711dafd4f785","Type":"ContainerDied","Data":"05af8c0b0757ab5358234087dfa025f5fe7f5339bb238dd5260ff8748c56647d"} Dec 10 19:09:49 crc kubenswrapper[4894]: I1210 19:09:49.717400 4894 generic.go:334] "Generic (PLEG): container finished" podID="877cd716-19e7-461e-b218-711dafd4f785" containerID="79b097be8065049fd889ab2c96c6ad6eaa670ac58e84ca3d1c1b49ee09941f69" exitCode=0 Dec 10 19:09:49 crc kubenswrapper[4894]: I1210 19:09:49.717488 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" event={"ID":"877cd716-19e7-461e-b218-711dafd4f785","Type":"ContainerDied","Data":"79b097be8065049fd889ab2c96c6ad6eaa670ac58e84ca3d1c1b49ee09941f69"} Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.008253 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.096666 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-bundle\") pod \"877cd716-19e7-461e-b218-711dafd4f785\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.096894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-util\") pod \"877cd716-19e7-461e-b218-711dafd4f785\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.097044 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftnlf\" (UniqueName: \"kubernetes.io/projected/877cd716-19e7-461e-b218-711dafd4f785-kube-api-access-ftnlf\") pod \"877cd716-19e7-461e-b218-711dafd4f785\" (UID: \"877cd716-19e7-461e-b218-711dafd4f785\") " Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.098526 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-bundle" (OuterVolumeSpecName: "bundle") pod "877cd716-19e7-461e-b218-711dafd4f785" (UID: "877cd716-19e7-461e-b218-711dafd4f785"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.103230 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877cd716-19e7-461e-b218-711dafd4f785-kube-api-access-ftnlf" (OuterVolumeSpecName: "kube-api-access-ftnlf") pod "877cd716-19e7-461e-b218-711dafd4f785" (UID: "877cd716-19e7-461e-b218-711dafd4f785"). InnerVolumeSpecName "kube-api-access-ftnlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.109145 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-util" (OuterVolumeSpecName: "util") pod "877cd716-19e7-461e-b218-711dafd4f785" (UID: "877cd716-19e7-461e-b218-711dafd4f785"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.198713 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.198775 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftnlf\" (UniqueName: \"kubernetes.io/projected/877cd716-19e7-461e-b218-711dafd4f785-kube-api-access-ftnlf\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.198800 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/877cd716-19e7-461e-b218-711dafd4f785-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.736466 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" event={"ID":"877cd716-19e7-461e-b218-711dafd4f785","Type":"ContainerDied","Data":"529a2c3a155b3499b6f62420267d07261f39313c165689e688e21edd75ce9d1b"} Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.736519 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="529a2c3a155b3499b6f62420267d07261f39313c165689e688e21edd75ce9d1b" Dec 10 19:09:51 crc kubenswrapper[4894]: I1210 19:09:51.736621 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.849717 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh"] Dec 10 19:09:56 crc kubenswrapper[4894]: E1210 19:09:56.850515 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="extract" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.850533 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="extract" Dec 10 19:09:56 crc kubenswrapper[4894]: E1210 19:09:56.850553 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="util" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.850561 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="util" Dec 10 19:09:56 crc kubenswrapper[4894]: E1210 19:09:56.850577 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="pull" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.850585 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="pull" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.850731 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="877cd716-19e7-461e-b218-711dafd4f785" containerName="extract" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.851351 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.853503 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.854065 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-22cq6" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.862520 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.886073 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbqs6\" (UniqueName: \"kubernetes.io/projected/c7d3b102-8456-4801-944d-7fb2b90cafc4-kube-api-access-tbqs6\") pod \"nmstate-operator-5b5b58f5c8-t5pwh\" (UID: \"c7d3b102-8456-4801-944d-7fb2b90cafc4\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.913971 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh"] Dec 10 19:09:56 crc kubenswrapper[4894]: I1210 19:09:56.990662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbqs6\" (UniqueName: \"kubernetes.io/projected/c7d3b102-8456-4801-944d-7fb2b90cafc4-kube-api-access-tbqs6\") pod \"nmstate-operator-5b5b58f5c8-t5pwh\" (UID: \"c7d3b102-8456-4801-944d-7fb2b90cafc4\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" Dec 10 19:09:57 crc kubenswrapper[4894]: I1210 19:09:57.013110 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbqs6\" (UniqueName: \"kubernetes.io/projected/c7d3b102-8456-4801-944d-7fb2b90cafc4-kube-api-access-tbqs6\") pod \"nmstate-operator-5b5b58f5c8-t5pwh\" (UID: \"c7d3b102-8456-4801-944d-7fb2b90cafc4\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" Dec 10 19:09:57 crc kubenswrapper[4894]: I1210 19:09:57.165890 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" Dec 10 19:09:57 crc kubenswrapper[4894]: I1210 19:09:57.393778 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh"] Dec 10 19:09:57 crc kubenswrapper[4894]: W1210 19:09:57.405955 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7d3b102_8456_4801_944d_7fb2b90cafc4.slice/crio-5f2db3371b84b5d8c810677adcd6c4ac929d436e520346be348b12250a7e6d56 WatchSource:0}: Error finding container 5f2db3371b84b5d8c810677adcd6c4ac929d436e520346be348b12250a7e6d56: Status 404 returned error can't find the container with id 5f2db3371b84b5d8c810677adcd6c4ac929d436e520346be348b12250a7e6d56 Dec 10 19:09:57 crc kubenswrapper[4894]: I1210 19:09:57.776213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" event={"ID":"c7d3b102-8456-4801-944d-7fb2b90cafc4","Type":"ContainerStarted","Data":"5f2db3371b84b5d8c810677adcd6c4ac929d436e520346be348b12250a7e6d56"} Dec 10 19:10:00 crc kubenswrapper[4894]: I1210 19:10:00.799543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" event={"ID":"c7d3b102-8456-4801-944d-7fb2b90cafc4","Type":"ContainerStarted","Data":"34c94b8fb774bdf9fa2c54f24e0a74ec52e67b3681096fc47aa74c0e5045f060"} Dec 10 19:10:00 crc kubenswrapper[4894]: I1210 19:10:00.825780 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t5pwh" podStartSLOduration=2.206128259 podStartE2EDuration="4.825755222s" podCreationTimestamp="2025-12-10 19:09:56 +0000 UTC" firstStartedPulling="2025-12-10 19:09:57.408731049 +0000 UTC m=+880.203578851" lastFinishedPulling="2025-12-10 19:10:00.028358022 +0000 UTC m=+882.823205814" observedRunningTime="2025-12-10 19:10:00.823457866 +0000 UTC m=+883.618305698" watchObservedRunningTime="2025-12-10 19:10:00.825755222 +0000 UTC m=+883.620603044" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.000685 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-g868p"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.001837 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.003779 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jksfb" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.016787 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-g868p"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.023393 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.024104 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.025991 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.051369 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.061910 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-knvqx"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.062684 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.114673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96zlt\" (UniqueName: \"kubernetes.io/projected/ab35ae17-b6ef-49e6-8cdb-625e03d9890d-kube-api-access-96zlt\") pod \"nmstate-metrics-7f946cbc9-g868p\" (UID: \"ab35ae17-b6ef-49e6-8cdb-625e03d9890d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.114737 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds879\" (UniqueName: \"kubernetes.io/projected/be4369f3-8902-40ee-9508-45698d96d6af-kube-api-access-ds879\") pod \"nmstate-webhook-5f6d4c5ccb-rh2pw\" (UID: \"be4369f3-8902-40ee-9508-45698d96d6af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.114759 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/be4369f3-8902-40ee-9508-45698d96d6af-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rh2pw\" (UID: \"be4369f3-8902-40ee-9508-45698d96d6af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.170224 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.171062 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.173588 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.173632 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-prprs" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.175265 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.182709 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmkt6\" (UniqueName: \"kubernetes.io/projected/f3832f69-624a-4be5-a69a-75e1d3200af4-kube-api-access-vmkt6\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216543 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-ovs-socket\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216591 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96zlt\" (UniqueName: \"kubernetes.io/projected/ab35ae17-b6ef-49e6-8cdb-625e03d9890d-kube-api-access-96zlt\") pod \"nmstate-metrics-7f946cbc9-g868p\" (UID: \"ab35ae17-b6ef-49e6-8cdb-625e03d9890d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds879\" (UniqueName: \"kubernetes.io/projected/be4369f3-8902-40ee-9508-45698d96d6af-kube-api-access-ds879\") pod \"nmstate-webhook-5f6d4c5ccb-rh2pw\" (UID: \"be4369f3-8902-40ee-9508-45698d96d6af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216634 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/be4369f3-8902-40ee-9508-45698d96d6af-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rh2pw\" (UID: \"be4369f3-8902-40ee-9508-45698d96d6af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216653 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-nmstate-lock\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.216673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-dbus-socket\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.236030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/be4369f3-8902-40ee-9508-45698d96d6af-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rh2pw\" (UID: \"be4369f3-8902-40ee-9508-45698d96d6af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.238690 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds879\" (UniqueName: \"kubernetes.io/projected/be4369f3-8902-40ee-9508-45698d96d6af-kube-api-access-ds879\") pod \"nmstate-webhook-5f6d4c5ccb-rh2pw\" (UID: \"be4369f3-8902-40ee-9508-45698d96d6af\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.240551 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96zlt\" (UniqueName: \"kubernetes.io/projected/ab35ae17-b6ef-49e6-8cdb-625e03d9890d-kube-api-access-96zlt\") pod \"nmstate-metrics-7f946cbc9-g868p\" (UID: \"ab35ae17-b6ef-49e6-8cdb-625e03d9890d\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-ovs-socket\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318178 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfkl4\" (UniqueName: \"kubernetes.io/projected/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-kube-api-access-bfkl4\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318231 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-nmstate-lock\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318224 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-ovs-socket\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318314 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-dbus-socket\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318339 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-nmstate-lock\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318373 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318520 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmkt6\" (UniqueName: \"kubernetes.io/projected/f3832f69-624a-4be5-a69a-75e1d3200af4-kube-api-access-vmkt6\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318570 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f3832f69-624a-4be5-a69a-75e1d3200af4-dbus-socket\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.318585 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.328624 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.340446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmkt6\" (UniqueName: \"kubernetes.io/projected/f3832f69-624a-4be5-a69a-75e1d3200af4-kube-api-access-vmkt6\") pod \"nmstate-handler-knvqx\" (UID: \"f3832f69-624a-4be5-a69a-75e1d3200af4\") " pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.342332 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.352786 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-67fdc9478d-xp49f"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.353725 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.370727 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67fdc9478d-xp49f"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.386233 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.420048 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.420112 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfkl4\" (UniqueName: \"kubernetes.io/projected/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-kube-api-access-bfkl4\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.420204 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.421092 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.426343 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: W1210 19:10:06.436355 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3832f69_624a_4be5_a69a_75e1d3200af4.slice/crio-359b5b9915123c2f141df9eaa1fe9f0231685cec54613a49a77aba6a2c4ef762 WatchSource:0}: Error finding container 359b5b9915123c2f141df9eaa1fe9f0231685cec54613a49a77aba6a2c4ef762: Status 404 returned error can't find the container with id 359b5b9915123c2f141df9eaa1fe9f0231685cec54613a49a77aba6a2c4ef762 Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.439911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfkl4\" (UniqueName: \"kubernetes.io/projected/5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02-kube-api-access-bfkl4\") pod \"nmstate-console-plugin-7fbb5f6569-qcks8\" (UID: \"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.484777 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.521769 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-oauth-serving-cert\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.521811 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-trusted-ca-bundle\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.521830 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab5cd659-60db-427f-b926-51602a64da5b-console-oauth-config\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.521873 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmz9z\" (UniqueName: \"kubernetes.io/projected/ab5cd659-60db-427f-b926-51602a64da5b-kube-api-access-qmz9z\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.521896 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-console-config\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.522031 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5cd659-60db-427f-b926-51602a64da5b-console-serving-cert\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.522153 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-service-ca\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.594056 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-g868p"] Dec 10 19:10:06 crc kubenswrapper[4894]: W1210 19:10:06.607753 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab35ae17_b6ef_49e6_8cdb_625e03d9890d.slice/crio-fd868b0bfebefe079b62a2d9f3c5f14ebf7151102286e83e0390db236c1cf463 WatchSource:0}: Error finding container fd868b0bfebefe079b62a2d9f3c5f14ebf7151102286e83e0390db236c1cf463: Status 404 returned error can't find the container with id fd868b0bfebefe079b62a2d9f3c5f14ebf7151102286e83e0390db236c1cf463 Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmz9z\" (UniqueName: \"kubernetes.io/projected/ab5cd659-60db-427f-b926-51602a64da5b-kube-api-access-qmz9z\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-console-config\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5cd659-60db-427f-b926-51602a64da5b-console-serving-cert\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624823 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-service-ca\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-oauth-serving-cert\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624955 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-trusted-ca-bundle\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.624979 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab5cd659-60db-427f-b926-51602a64da5b-console-oauth-config\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.626062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-service-ca\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.626117 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-console-config\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.626167 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-oauth-serving-cert\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.626524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab5cd659-60db-427f-b926-51602a64da5b-trusted-ca-bundle\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.631044 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab5cd659-60db-427f-b926-51602a64da5b-console-oauth-config\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.631565 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5cd659-60db-427f-b926-51602a64da5b-console-serving-cert\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.632054 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw"] Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.644824 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmz9z\" (UniqueName: \"kubernetes.io/projected/ab5cd659-60db-427f-b926-51602a64da5b-kube-api-access-qmz9z\") pod \"console-67fdc9478d-xp49f\" (UID: \"ab5cd659-60db-427f-b926-51602a64da5b\") " pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.692114 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8"] Dec 10 19:10:06 crc kubenswrapper[4894]: W1210 19:10:06.699283 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d0c2c6f_f293_48cd_a22b_d77fc2fd9b02.slice/crio-0fda983aeb4f0c8ae2222472f774d3542977b9542fa79bd9a32ade09361fa472 WatchSource:0}: Error finding container 0fda983aeb4f0c8ae2222472f774d3542977b9542fa79bd9a32ade09361fa472: Status 404 returned error can't find the container with id 0fda983aeb4f0c8ae2222472f774d3542977b9542fa79bd9a32ade09361fa472 Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.737688 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.836537 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" event={"ID":"ab35ae17-b6ef-49e6-8cdb-625e03d9890d","Type":"ContainerStarted","Data":"fd868b0bfebefe079b62a2d9f3c5f14ebf7151102286e83e0390db236c1cf463"} Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.837730 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" event={"ID":"be4369f3-8902-40ee-9508-45698d96d6af","Type":"ContainerStarted","Data":"697bfe7d1c2bf6180065dc05e6d7068476ebccccbf6ff801b4fa87b26b0cc0a0"} Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.838594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" event={"ID":"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02","Type":"ContainerStarted","Data":"0fda983aeb4f0c8ae2222472f774d3542977b9542fa79bd9a32ade09361fa472"} Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.839305 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-knvqx" event={"ID":"f3832f69-624a-4be5-a69a-75e1d3200af4","Type":"ContainerStarted","Data":"359b5b9915123c2f141df9eaa1fe9f0231685cec54613a49a77aba6a2c4ef762"} Dec 10 19:10:06 crc kubenswrapper[4894]: I1210 19:10:06.935151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67fdc9478d-xp49f"] Dec 10 19:10:06 crc kubenswrapper[4894]: W1210 19:10:06.940894 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab5cd659_60db_427f_b926_51602a64da5b.slice/crio-6d4b8587316128ae0ff7f722ebe3330df641eb3df1c75780c3fb9f09db67fe5a WatchSource:0}: Error finding container 6d4b8587316128ae0ff7f722ebe3330df641eb3df1c75780c3fb9f09db67fe5a: Status 404 returned error can't find the container with id 6d4b8587316128ae0ff7f722ebe3330df641eb3df1c75780c3fb9f09db67fe5a Dec 10 19:10:07 crc kubenswrapper[4894]: I1210 19:10:07.846216 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67fdc9478d-xp49f" event={"ID":"ab5cd659-60db-427f-b926-51602a64da5b","Type":"ContainerStarted","Data":"85b112abdccd30bd7432734692dc50d26f780fcd1e68a8d3b27966de94d4095c"} Dec 10 19:10:07 crc kubenswrapper[4894]: I1210 19:10:07.846798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67fdc9478d-xp49f" event={"ID":"ab5cd659-60db-427f-b926-51602a64da5b","Type":"ContainerStarted","Data":"6d4b8587316128ae0ff7f722ebe3330df641eb3df1c75780c3fb9f09db67fe5a"} Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.873323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" event={"ID":"ab35ae17-b6ef-49e6-8cdb-625e03d9890d","Type":"ContainerStarted","Data":"4c1ad409388d322e8b650efdac11ac4320becf57bfbd784fe3db2304ee529cb9"} Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.875076 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" event={"ID":"be4369f3-8902-40ee-9508-45698d96d6af","Type":"ContainerStarted","Data":"407663680ad201d3392ca337457d44d7bdb132a970945b70718b4a090898bed7"} Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.875482 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.877467 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" event={"ID":"5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02","Type":"ContainerStarted","Data":"d619c27ce6d4536646d5880fc69b43c290ada3748ff1854e6885f3143301f3c2"} Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.885275 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-knvqx" event={"ID":"f3832f69-624a-4be5-a69a-75e1d3200af4","Type":"ContainerStarted","Data":"459d95ceb5f6fb79a9bbed1a29512b4fb07cac5f504f12fdb2b66969af6cfb63"} Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.885503 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.897097 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" podStartSLOduration=1.620054454 podStartE2EDuration="4.897076356s" podCreationTimestamp="2025-12-10 19:10:06 +0000 UTC" firstStartedPulling="2025-12-10 19:10:06.639692501 +0000 UTC m=+889.434540293" lastFinishedPulling="2025-12-10 19:10:09.916714413 +0000 UTC m=+892.711562195" observedRunningTime="2025-12-10 19:10:10.895491537 +0000 UTC m=+893.690339349" watchObservedRunningTime="2025-12-10 19:10:10.897076356 +0000 UTC m=+893.691924158" Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.901342 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-67fdc9478d-xp49f" podStartSLOduration=4.901323401 podStartE2EDuration="4.901323401s" podCreationTimestamp="2025-12-10 19:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:10:07.882098285 +0000 UTC m=+890.676946077" watchObservedRunningTime="2025-12-10 19:10:10.901323401 +0000 UTC m=+893.696171213" Dec 10 19:10:10 crc kubenswrapper[4894]: I1210 19:10:10.961819 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-knvqx" podStartSLOduration=1.458807983 podStartE2EDuration="4.961797343s" podCreationTimestamp="2025-12-10 19:10:06 +0000 UTC" firstStartedPulling="2025-12-10 19:10:06.438960181 +0000 UTC m=+889.233807973" lastFinishedPulling="2025-12-10 19:10:09.941949531 +0000 UTC m=+892.736797333" observedRunningTime="2025-12-10 19:10:10.948968818 +0000 UTC m=+893.743816610" watchObservedRunningTime="2025-12-10 19:10:10.961797343 +0000 UTC m=+893.756645135" Dec 10 19:10:12 crc kubenswrapper[4894]: I1210 19:10:12.903962 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" event={"ID":"ab35ae17-b6ef-49e6-8cdb-625e03d9890d","Type":"ContainerStarted","Data":"35699bc86d98f1760fc079859b205b108cfadd24cab4da8aa5d0dde8a0a27d4e"} Dec 10 19:10:12 crc kubenswrapper[4894]: I1210 19:10:12.940174 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g868p" podStartSLOduration=2.389361471 podStartE2EDuration="7.940153492s" podCreationTimestamp="2025-12-10 19:10:05 +0000 UTC" firstStartedPulling="2025-12-10 19:10:06.60948171 +0000 UTC m=+889.404329502" lastFinishedPulling="2025-12-10 19:10:12.160273731 +0000 UTC m=+894.955121523" observedRunningTime="2025-12-10 19:10:12.935502507 +0000 UTC m=+895.730350289" watchObservedRunningTime="2025-12-10 19:10:12.940153492 +0000 UTC m=+895.735001294" Dec 10 19:10:12 crc kubenswrapper[4894]: I1210 19:10:12.941043 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qcks8" podStartSLOduration=3.725331342 podStartE2EDuration="6.941037043s" podCreationTimestamp="2025-12-10 19:10:06 +0000 UTC" firstStartedPulling="2025-12-10 19:10:06.700751236 +0000 UTC m=+889.495599028" lastFinishedPulling="2025-12-10 19:10:09.916456927 +0000 UTC m=+892.711304729" observedRunningTime="2025-12-10 19:10:10.97111217 +0000 UTC m=+893.765959962" watchObservedRunningTime="2025-12-10 19:10:12.941037043 +0000 UTC m=+895.735884845" Dec 10 19:10:16 crc kubenswrapper[4894]: I1210 19:10:16.430018 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-knvqx" Dec 10 19:10:16 crc kubenswrapper[4894]: I1210 19:10:16.738335 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:16 crc kubenswrapper[4894]: I1210 19:10:16.738407 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:16 crc kubenswrapper[4894]: I1210 19:10:16.744457 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:16 crc kubenswrapper[4894]: I1210 19:10:16.942675 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-67fdc9478d-xp49f" Dec 10 19:10:17 crc kubenswrapper[4894]: I1210 19:10:17.013394 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vsms7"] Dec 10 19:10:26 crc kubenswrapper[4894]: I1210 19:10:26.353161 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rh2pw" Dec 10 19:10:36 crc kubenswrapper[4894]: I1210 19:10:36.936531 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s2hmq"] Dec 10 19:10:36 crc kubenswrapper[4894]: I1210 19:10:36.938300 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:36 crc kubenswrapper[4894]: I1210 19:10:36.948502 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2hmq"] Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.040905 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-utilities\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.041052 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-catalog-content\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.041105 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqqhp\" (UniqueName: \"kubernetes.io/projected/2f9a4969-6b38-49b6-832c-5d747feb1bc6-kube-api-access-bqqhp\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.142075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-utilities\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.142121 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-catalog-content\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.142140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqqhp\" (UniqueName: \"kubernetes.io/projected/2f9a4969-6b38-49b6-832c-5d747feb1bc6-kube-api-access-bqqhp\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.143209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-catalog-content\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.143393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-utilities\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.164139 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqqhp\" (UniqueName: \"kubernetes.io/projected/2f9a4969-6b38-49b6-832c-5d747feb1bc6-kube-api-access-bqqhp\") pod \"redhat-marketplace-s2hmq\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.277592 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:37 crc kubenswrapper[4894]: I1210 19:10:37.721043 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2hmq"] Dec 10 19:10:38 crc kubenswrapper[4894]: I1210 19:10:38.086072 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerID="14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98" exitCode=0 Dec 10 19:10:38 crc kubenswrapper[4894]: I1210 19:10:38.086125 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2hmq" event={"ID":"2f9a4969-6b38-49b6-832c-5d747feb1bc6","Type":"ContainerDied","Data":"14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98"} Dec 10 19:10:38 crc kubenswrapper[4894]: I1210 19:10:38.086516 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2hmq" event={"ID":"2f9a4969-6b38-49b6-832c-5d747feb1bc6","Type":"ContainerStarted","Data":"1229449cd7781df9bfa881515ae31b77c1a7f8d7cc50b0b1bd7d4f96ac3e8a30"} Dec 10 19:10:40 crc kubenswrapper[4894]: I1210 19:10:40.104553 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerID="b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98" exitCode=0 Dec 10 19:10:40 crc kubenswrapper[4894]: I1210 19:10:40.104664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2hmq" event={"ID":"2f9a4969-6b38-49b6-832c-5d747feb1bc6","Type":"ContainerDied","Data":"b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98"} Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.111822 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2hmq" event={"ID":"2f9a4969-6b38-49b6-832c-5d747feb1bc6","Type":"ContainerStarted","Data":"5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659"} Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.130919 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s2hmq" podStartSLOduration=2.628758504 podStartE2EDuration="5.130882597s" podCreationTimestamp="2025-12-10 19:10:36 +0000 UTC" firstStartedPulling="2025-12-10 19:10:38.08832155 +0000 UTC m=+920.883169342" lastFinishedPulling="2025-12-10 19:10:40.590445653 +0000 UTC m=+923.385293435" observedRunningTime="2025-12-10 19:10:41.126002157 +0000 UTC m=+923.920849949" watchObservedRunningTime="2025-12-10 19:10:41.130882597 +0000 UTC m=+923.925730409" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.574718 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2"] Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.576177 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.578380 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.650895 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2"] Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.712642 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.713023 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cglh9\" (UniqueName: \"kubernetes.io/projected/e20fd98f-3383-4462-9894-ab6c867c0064-kube-api-access-cglh9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.713122 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.814525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.814634 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.814668 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cglh9\" (UniqueName: \"kubernetes.io/projected/e20fd98f-3383-4462-9894-ab6c867c0064-kube-api-access-cglh9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.815307 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.815425 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.833032 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cglh9\" (UniqueName: \"kubernetes.io/projected/e20fd98f-3383-4462-9894-ab6c867c0064-kube-api-access-cglh9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:41 crc kubenswrapper[4894]: I1210 19:10:41.947169 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.086022 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vsms7" podUID="139baf99-78f8-41f5-8c3c-2d37c42edd47" containerName="console" containerID="cri-o://5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221" gracePeriod=15 Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.160854 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2"] Dec 10 19:10:42 crc kubenswrapper[4894]: W1210 19:10:42.191656 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode20fd98f_3383_4462_9894_ab6c867c0064.slice/crio-63a9edaa0b54904570a5d9f7895b32c8ae08e7e82ce3986ae05c61e3dddf5746 WatchSource:0}: Error finding container 63a9edaa0b54904570a5d9f7895b32c8ae08e7e82ce3986ae05c61e3dddf5746: Status 404 returned error can't find the container with id 63a9edaa0b54904570a5d9f7895b32c8ae08e7e82ce3986ae05c61e3dddf5746 Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.464804 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vsms7_139baf99-78f8-41f5-8c3c-2d37c42edd47/console/0.log" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.465229 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.624975 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9kck\" (UniqueName: \"kubernetes.io/projected/139baf99-78f8-41f5-8c3c-2d37c42edd47-kube-api-access-w9kck\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.625019 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-config\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.625109 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-oauth-serving-cert\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.625131 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-oauth-config\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.625177 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-serving-cert\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.625201 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.625220 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-service-ca\") pod \"139baf99-78f8-41f5-8c3c-2d37c42edd47\" (UID: \"139baf99-78f8-41f5-8c3c-2d37c42edd47\") " Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.626248 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.626271 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-service-ca" (OuterVolumeSpecName: "service-ca") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.626292 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.626553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-config" (OuterVolumeSpecName: "console-config") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.630202 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.630530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139baf99-78f8-41f5-8c3c-2d37c42edd47-kube-api-access-w9kck" (OuterVolumeSpecName: "kube-api-access-w9kck") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "kube-api-access-w9kck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.631621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "139baf99-78f8-41f5-8c3c-2d37c42edd47" (UID: "139baf99-78f8-41f5-8c3c-2d37c42edd47"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726806 4894 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726875 4894 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726897 4894 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726915 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726933 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726954 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9kck\" (UniqueName: \"kubernetes.io/projected/139baf99-78f8-41f5-8c3c-2d37c42edd47-kube-api-access-w9kck\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.726973 4894 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/139baf99-78f8-41f5-8c3c-2d37c42edd47-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.907670 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tj2lx"] Dec 10 19:10:42 crc kubenswrapper[4894]: E1210 19:10:42.907950 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139baf99-78f8-41f5-8c3c-2d37c42edd47" containerName="console" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.907964 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="139baf99-78f8-41f5-8c3c-2d37c42edd47" containerName="console" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.908116 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="139baf99-78f8-41f5-8c3c-2d37c42edd47" containerName="console" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.909055 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:42 crc kubenswrapper[4894]: I1210 19:10:42.936683 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tj2lx"] Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.030823 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5ppl\" (UniqueName: \"kubernetes.io/projected/29a03790-f1bb-4d04-ac00-7c9bd894300d-kube-api-access-v5ppl\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.031159 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-catalog-content\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.031651 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-utilities\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.133300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5ppl\" (UniqueName: \"kubernetes.io/projected/29a03790-f1bb-4d04-ac00-7c9bd894300d-kube-api-access-v5ppl\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.133361 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-catalog-content\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.133438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-utilities\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.133918 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-utilities\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.134078 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-catalog-content\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.142452 4894 generic.go:334] "Generic (PLEG): container finished" podID="e20fd98f-3383-4462-9894-ab6c867c0064" containerID="50819c132d2190b1f481cb7185c1a946f84cb3a4daf105596247b20217b9bf02" exitCode=0 Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.142507 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" event={"ID":"e20fd98f-3383-4462-9894-ab6c867c0064","Type":"ContainerDied","Data":"50819c132d2190b1f481cb7185c1a946f84cb3a4daf105596247b20217b9bf02"} Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.142565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" event={"ID":"e20fd98f-3383-4462-9894-ab6c867c0064","Type":"ContainerStarted","Data":"63a9edaa0b54904570a5d9f7895b32c8ae08e7e82ce3986ae05c61e3dddf5746"} Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.145556 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vsms7_139baf99-78f8-41f5-8c3c-2d37c42edd47/console/0.log" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.145596 4894 generic.go:334] "Generic (PLEG): container finished" podID="139baf99-78f8-41f5-8c3c-2d37c42edd47" containerID="5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221" exitCode=2 Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.145619 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vsms7" event={"ID":"139baf99-78f8-41f5-8c3c-2d37c42edd47","Type":"ContainerDied","Data":"5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221"} Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.145643 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vsms7" event={"ID":"139baf99-78f8-41f5-8c3c-2d37c42edd47","Type":"ContainerDied","Data":"261ce7371bbe067233bc7f45cc17e5e9464c73055de09905b019ba2f16088d27"} Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.145659 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vsms7" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.145674 4894 scope.go:117] "RemoveContainer" containerID="5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.152779 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5ppl\" (UniqueName: \"kubernetes.io/projected/29a03790-f1bb-4d04-ac00-7c9bd894300d-kube-api-access-v5ppl\") pod \"community-operators-tj2lx\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.193143 4894 scope.go:117] "RemoveContainer" containerID="5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221" Dec 10 19:10:43 crc kubenswrapper[4894]: E1210 19:10:43.194050 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221\": container with ID starting with 5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221 not found: ID does not exist" containerID="5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.194089 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221"} err="failed to get container status \"5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221\": rpc error: code = NotFound desc = could not find container \"5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221\": container with ID starting with 5538b380ccfa65be343233930903f9fb617a37f9f2d411b730f4243e5ded4221 not found: ID does not exist" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.202067 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vsms7"] Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.206602 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vsms7"] Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.234818 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.550060 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="139baf99-78f8-41f5-8c3c-2d37c42edd47" path="/var/lib/kubelet/pods/139baf99-78f8-41f5-8c3c-2d37c42edd47/volumes" Dec 10 19:10:43 crc kubenswrapper[4894]: I1210 19:10:43.694910 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tj2lx"] Dec 10 19:10:44 crc kubenswrapper[4894]: I1210 19:10:44.159941 4894 generic.go:334] "Generic (PLEG): container finished" podID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerID="e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013" exitCode=0 Dec 10 19:10:44 crc kubenswrapper[4894]: I1210 19:10:44.160011 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tj2lx" event={"ID":"29a03790-f1bb-4d04-ac00-7c9bd894300d","Type":"ContainerDied","Data":"e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013"} Dec 10 19:10:44 crc kubenswrapper[4894]: I1210 19:10:44.160270 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tj2lx" event={"ID":"29a03790-f1bb-4d04-ac00-7c9bd894300d","Type":"ContainerStarted","Data":"db17ce53d5d2d488dba7d66739c63613524aef5724625b866b67ce1e3cf8acfd"} Dec 10 19:10:45 crc kubenswrapper[4894]: I1210 19:10:45.171670 4894 generic.go:334] "Generic (PLEG): container finished" podID="e20fd98f-3383-4462-9894-ab6c867c0064" containerID="4a0a2980fe52c4e2aeb4d79b6f8547546530a5498bf92c5e9413f5d3b930465a" exitCode=0 Dec 10 19:10:45 crc kubenswrapper[4894]: I1210 19:10:45.171709 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" event={"ID":"e20fd98f-3383-4462-9894-ab6c867c0064","Type":"ContainerDied","Data":"4a0a2980fe52c4e2aeb4d79b6f8547546530a5498bf92c5e9413f5d3b930465a"} Dec 10 19:10:46 crc kubenswrapper[4894]: I1210 19:10:46.177309 4894 generic.go:334] "Generic (PLEG): container finished" podID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerID="aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377" exitCode=0 Dec 10 19:10:46 crc kubenswrapper[4894]: I1210 19:10:46.177363 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tj2lx" event={"ID":"29a03790-f1bb-4d04-ac00-7c9bd894300d","Type":"ContainerDied","Data":"aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377"} Dec 10 19:10:46 crc kubenswrapper[4894]: I1210 19:10:46.182337 4894 generic.go:334] "Generic (PLEG): container finished" podID="e20fd98f-3383-4462-9894-ab6c867c0064" containerID="a66a8620224bd871f8a04377dfdc7c53748203473b5d58feb3ebb5723ed7ac9e" exitCode=0 Dec 10 19:10:46 crc kubenswrapper[4894]: I1210 19:10:46.182361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" event={"ID":"e20fd98f-3383-4462-9894-ab6c867c0064","Type":"ContainerDied","Data":"a66a8620224bd871f8a04377dfdc7c53748203473b5d58feb3ebb5723ed7ac9e"} Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.194646 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tj2lx" event={"ID":"29a03790-f1bb-4d04-ac00-7c9bd894300d","Type":"ContainerStarted","Data":"707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1"} Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.278531 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.278591 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.332410 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.361117 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tj2lx" podStartSLOduration=2.6697211149999998 podStartE2EDuration="5.361101006s" podCreationTimestamp="2025-12-10 19:10:42 +0000 UTC" firstStartedPulling="2025-12-10 19:10:44.1627153 +0000 UTC m=+926.957563102" lastFinishedPulling="2025-12-10 19:10:46.854095161 +0000 UTC m=+929.648942993" observedRunningTime="2025-12-10 19:10:47.231040989 +0000 UTC m=+930.025888791" watchObservedRunningTime="2025-12-10 19:10:47.361101006 +0000 UTC m=+930.155948808" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.460673 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.593653 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-util\") pod \"e20fd98f-3383-4462-9894-ab6c867c0064\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.593754 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cglh9\" (UniqueName: \"kubernetes.io/projected/e20fd98f-3383-4462-9894-ab6c867c0064-kube-api-access-cglh9\") pod \"e20fd98f-3383-4462-9894-ab6c867c0064\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.593800 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-bundle\") pod \"e20fd98f-3383-4462-9894-ab6c867c0064\" (UID: \"e20fd98f-3383-4462-9894-ab6c867c0064\") " Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.595324 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-bundle" (OuterVolumeSpecName: "bundle") pod "e20fd98f-3383-4462-9894-ab6c867c0064" (UID: "e20fd98f-3383-4462-9894-ab6c867c0064"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.605232 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e20fd98f-3383-4462-9894-ab6c867c0064-kube-api-access-cglh9" (OuterVolumeSpecName: "kube-api-access-cglh9") pod "e20fd98f-3383-4462-9894-ab6c867c0064" (UID: "e20fd98f-3383-4462-9894-ab6c867c0064"). InnerVolumeSpecName "kube-api-access-cglh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.625892 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-util" (OuterVolumeSpecName: "util") pod "e20fd98f-3383-4462-9894-ab6c867c0064" (UID: "e20fd98f-3383-4462-9894-ab6c867c0064"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.698359 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cglh9\" (UniqueName: \"kubernetes.io/projected/e20fd98f-3383-4462-9894-ab6c867c0064-kube-api-access-cglh9\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.698804 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:47 crc kubenswrapper[4894]: I1210 19:10:47.698832 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e20fd98f-3383-4462-9894-ab6c867c0064-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:48 crc kubenswrapper[4894]: I1210 19:10:48.204752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" event={"ID":"e20fd98f-3383-4462-9894-ab6c867c0064","Type":"ContainerDied","Data":"63a9edaa0b54904570a5d9f7895b32c8ae08e7e82ce3986ae05c61e3dddf5746"} Dec 10 19:10:48 crc kubenswrapper[4894]: I1210 19:10:48.204818 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a9edaa0b54904570a5d9f7895b32c8ae08e7e82ce3986ae05c61e3dddf5746" Dec 10 19:10:48 crc kubenswrapper[4894]: I1210 19:10:48.204834 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2" Dec 10 19:10:48 crc kubenswrapper[4894]: I1210 19:10:48.262533 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:49 crc kubenswrapper[4894]: I1210 19:10:49.502657 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2hmq"] Dec 10 19:10:49 crc kubenswrapper[4894]: I1210 19:10:49.579796 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:10:49 crc kubenswrapper[4894]: I1210 19:10:49.579931 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.222338 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s2hmq" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="registry-server" containerID="cri-o://5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659" gracePeriod=2 Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.587131 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.744672 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-catalog-content\") pod \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.745131 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqqhp\" (UniqueName: \"kubernetes.io/projected/2f9a4969-6b38-49b6-832c-5d747feb1bc6-kube-api-access-bqqhp\") pod \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.745161 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-utilities\") pod \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\" (UID: \"2f9a4969-6b38-49b6-832c-5d747feb1bc6\") " Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.745972 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-utilities" (OuterVolumeSpecName: "utilities") pod "2f9a4969-6b38-49b6-832c-5d747feb1bc6" (UID: "2f9a4969-6b38-49b6-832c-5d747feb1bc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.750992 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f9a4969-6b38-49b6-832c-5d747feb1bc6-kube-api-access-bqqhp" (OuterVolumeSpecName: "kube-api-access-bqqhp") pod "2f9a4969-6b38-49b6-832c-5d747feb1bc6" (UID: "2f9a4969-6b38-49b6-832c-5d747feb1bc6"). InnerVolumeSpecName "kube-api-access-bqqhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.772315 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f9a4969-6b38-49b6-832c-5d747feb1bc6" (UID: "2f9a4969-6b38-49b6-832c-5d747feb1bc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.846028 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqqhp\" (UniqueName: \"kubernetes.io/projected/2f9a4969-6b38-49b6-832c-5d747feb1bc6-kube-api-access-bqqhp\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.846063 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:50 crc kubenswrapper[4894]: I1210 19:10:50.846074 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9a4969-6b38-49b6-832c-5d747feb1bc6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.233437 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerID="5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659" exitCode=0 Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.233485 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2hmq" event={"ID":"2f9a4969-6b38-49b6-832c-5d747feb1bc6","Type":"ContainerDied","Data":"5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659"} Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.233514 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2hmq" event={"ID":"2f9a4969-6b38-49b6-832c-5d747feb1bc6","Type":"ContainerDied","Data":"1229449cd7781df9bfa881515ae31b77c1a7f8d7cc50b0b1bd7d4f96ac3e8a30"} Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.233534 4894 scope.go:117] "RemoveContainer" containerID="5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.233660 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2hmq" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.268837 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2hmq"] Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.269677 4894 scope.go:117] "RemoveContainer" containerID="b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.275331 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2hmq"] Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.285488 4894 scope.go:117] "RemoveContainer" containerID="14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.312769 4894 scope.go:117] "RemoveContainer" containerID="5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659" Dec 10 19:10:51 crc kubenswrapper[4894]: E1210 19:10:51.313217 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659\": container with ID starting with 5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659 not found: ID does not exist" containerID="5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.313255 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659"} err="failed to get container status \"5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659\": rpc error: code = NotFound desc = could not find container \"5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659\": container with ID starting with 5e5199d63958e341f93b3df3687ebb3574c351672957eebb67b5b599a1833659 not found: ID does not exist" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.313283 4894 scope.go:117] "RemoveContainer" containerID="b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98" Dec 10 19:10:51 crc kubenswrapper[4894]: E1210 19:10:51.313625 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98\": container with ID starting with b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98 not found: ID does not exist" containerID="b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.313650 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98"} err="failed to get container status \"b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98\": rpc error: code = NotFound desc = could not find container \"b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98\": container with ID starting with b035bf3b30f011c42ed7a389e8860a8fab9f2eb1e4bedb0f626c111421ec3a98 not found: ID does not exist" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.313670 4894 scope.go:117] "RemoveContainer" containerID="14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98" Dec 10 19:10:51 crc kubenswrapper[4894]: E1210 19:10:51.314147 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98\": container with ID starting with 14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98 not found: ID does not exist" containerID="14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.314195 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98"} err="failed to get container status \"14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98\": rpc error: code = NotFound desc = could not find container \"14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98\": container with ID starting with 14032a060b33eb343387528558fad0a6f1156aa172196845e25da2b505ad5c98 not found: ID does not exist" Dec 10 19:10:51 crc kubenswrapper[4894]: I1210 19:10:51.549214 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" path="/var/lib/kubelet/pods/2f9a4969-6b38-49b6-832c-5d747feb1bc6/volumes" Dec 10 19:10:53 crc kubenswrapper[4894]: I1210 19:10:53.235678 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:53 crc kubenswrapper[4894]: I1210 19:10:53.237724 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:53 crc kubenswrapper[4894]: I1210 19:10:53.275467 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:53 crc kubenswrapper[4894]: I1210 19:10:53.327256 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:55 crc kubenswrapper[4894]: I1210 19:10:55.891354 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tj2lx"] Dec 10 19:10:55 crc kubenswrapper[4894]: I1210 19:10:55.891893 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tj2lx" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="registry-server" containerID="cri-o://707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1" gracePeriod=2 Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.810749 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.923175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-catalog-content\") pod \"29a03790-f1bb-4d04-ac00-7c9bd894300d\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.923310 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-utilities\") pod \"29a03790-f1bb-4d04-ac00-7c9bd894300d\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.923347 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5ppl\" (UniqueName: \"kubernetes.io/projected/29a03790-f1bb-4d04-ac00-7c9bd894300d-kube-api-access-v5ppl\") pod \"29a03790-f1bb-4d04-ac00-7c9bd894300d\" (UID: \"29a03790-f1bb-4d04-ac00-7c9bd894300d\") " Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.931987 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a03790-f1bb-4d04-ac00-7c9bd894300d-kube-api-access-v5ppl" (OuterVolumeSpecName: "kube-api-access-v5ppl") pod "29a03790-f1bb-4d04-ac00-7c9bd894300d" (UID: "29a03790-f1bb-4d04-ac00-7c9bd894300d"). InnerVolumeSpecName "kube-api-access-v5ppl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.933361 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-utilities" (OuterVolumeSpecName: "utilities") pod "29a03790-f1bb-4d04-ac00-7c9bd894300d" (UID: "29a03790-f1bb-4d04-ac00-7c9bd894300d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:56 crc kubenswrapper[4894]: I1210 19:10:56.990205 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29a03790-f1bb-4d04-ac00-7c9bd894300d" (UID: "29a03790-f1bb-4d04-ac00-7c9bd894300d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.024702 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.024739 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5ppl\" (UniqueName: \"kubernetes.io/projected/29a03790-f1bb-4d04-ac00-7c9bd894300d-kube-api-access-v5ppl\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.024752 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a03790-f1bb-4d04-ac00-7c9bd894300d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.272548 4894 generic.go:334] "Generic (PLEG): container finished" podID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerID="707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1" exitCode=0 Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.272592 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tj2lx" event={"ID":"29a03790-f1bb-4d04-ac00-7c9bd894300d","Type":"ContainerDied","Data":"707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1"} Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.272625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tj2lx" event={"ID":"29a03790-f1bb-4d04-ac00-7c9bd894300d","Type":"ContainerDied","Data":"db17ce53d5d2d488dba7d66739c63613524aef5724625b866b67ce1e3cf8acfd"} Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.272648 4894 scope.go:117] "RemoveContainer" containerID="707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.272712 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tj2lx" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.298521 4894 scope.go:117] "RemoveContainer" containerID="aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.326862 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tj2lx"] Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.337827 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tj2lx"] Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.338447 4894 scope.go:117] "RemoveContainer" containerID="e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.355566 4894 scope.go:117] "RemoveContainer" containerID="707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1" Dec 10 19:10:57 crc kubenswrapper[4894]: E1210 19:10:57.356600 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1\": container with ID starting with 707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1 not found: ID does not exist" containerID="707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.356711 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1"} err="failed to get container status \"707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1\": rpc error: code = NotFound desc = could not find container \"707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1\": container with ID starting with 707b6cb6f1f960ec7a1cb8bb2b1afb3be252544cf091808a269ee7d1ebb948d1 not found: ID does not exist" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.356813 4894 scope.go:117] "RemoveContainer" containerID="aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377" Dec 10 19:10:57 crc kubenswrapper[4894]: E1210 19:10:57.357763 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377\": container with ID starting with aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377 not found: ID does not exist" containerID="aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.357861 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377"} err="failed to get container status \"aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377\": rpc error: code = NotFound desc = could not find container \"aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377\": container with ID starting with aef448ca2a9bef8dccfccdc3cfd6ffdad799ee8c8c57756502a8e4616469c377 not found: ID does not exist" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.357927 4894 scope.go:117] "RemoveContainer" containerID="e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013" Dec 10 19:10:57 crc kubenswrapper[4894]: E1210 19:10:57.358274 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013\": container with ID starting with e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013 not found: ID does not exist" containerID="e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.358356 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013"} err="failed to get container status \"e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013\": rpc error: code = NotFound desc = could not find container \"e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013\": container with ID starting with e0d013fe15a9b691db01164296e20b1a10f6fa2b63ba7ea78d489325899d4013 not found: ID does not exist" Dec 10 19:10:57 crc kubenswrapper[4894]: I1210 19:10:57.551490 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" path="/var/lib/kubelet/pods/29a03790-f1bb-4d04-ac00-7c9bd894300d/volumes" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.469832 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8"] Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470344 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="registry-server" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470357 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="registry-server" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470364 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="pull" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470370 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="pull" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470381 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="extract-utilities" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470388 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="extract-utilities" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470397 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="extract-content" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470403 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="extract-content" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470417 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="extract-content" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470422 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="extract-content" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470430 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="extract-utilities" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470436 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="extract-utilities" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470443 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="registry-server" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470448 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="registry-server" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470456 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="util" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470461 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="util" Dec 10 19:10:59 crc kubenswrapper[4894]: E1210 19:10:59.470468 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="extract" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470474 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="extract" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470590 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a03790-f1bb-4d04-ac00-7c9bd894300d" containerName="registry-server" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470598 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20fd98f-3383-4462-9894-ab6c867c0064" containerName="extract" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.470606 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f9a4969-6b38-49b6-832c-5d747feb1bc6" containerName="registry-server" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.471010 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.473451 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.473648 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.474096 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.474231 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-q262t" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.474403 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.484498 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8"] Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.560610 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/25e24603-77a9-43cd-86c6-941b45ab2d4b-apiservice-cert\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.560675 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rj65\" (UniqueName: \"kubernetes.io/projected/25e24603-77a9-43cd-86c6-941b45ab2d4b-kube-api-access-4rj65\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.560776 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/25e24603-77a9-43cd-86c6-941b45ab2d4b-webhook-cert\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.662496 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/25e24603-77a9-43cd-86c6-941b45ab2d4b-webhook-cert\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.662558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/25e24603-77a9-43cd-86c6-941b45ab2d4b-apiservice-cert\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.662579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rj65\" (UniqueName: \"kubernetes.io/projected/25e24603-77a9-43cd-86c6-941b45ab2d4b-kube-api-access-4rj65\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.671598 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/25e24603-77a9-43cd-86c6-941b45ab2d4b-apiservice-cert\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.671611 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/25e24603-77a9-43cd-86c6-941b45ab2d4b-webhook-cert\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.678571 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rj65\" (UniqueName: \"kubernetes.io/projected/25e24603-77a9-43cd-86c6-941b45ab2d4b-kube-api-access-4rj65\") pod \"metallb-operator-controller-manager-8484c557fc-rnmp8\" (UID: \"25e24603-77a9-43cd-86c6-941b45ab2d4b\") " pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.718677 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8"] Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.719437 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.721780 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.721780 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.722234 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2lnqt" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.740676 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8"] Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.764074 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60aa603c-1807-48aa-80c0-fd56eedeff49-webhook-cert\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.764388 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r6f4\" (UniqueName: \"kubernetes.io/projected/60aa603c-1807-48aa-80c0-fd56eedeff49-kube-api-access-7r6f4\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.764472 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60aa603c-1807-48aa-80c0-fd56eedeff49-apiservice-cert\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.788015 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.867573 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60aa603c-1807-48aa-80c0-fd56eedeff49-webhook-cert\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.867604 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r6f4\" (UniqueName: \"kubernetes.io/projected/60aa603c-1807-48aa-80c0-fd56eedeff49-kube-api-access-7r6f4\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.867625 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60aa603c-1807-48aa-80c0-fd56eedeff49-apiservice-cert\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.879342 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60aa603c-1807-48aa-80c0-fd56eedeff49-webhook-cert\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.881559 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60aa603c-1807-48aa-80c0-fd56eedeff49-apiservice-cert\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:10:59 crc kubenswrapper[4894]: I1210 19:10:59.888040 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r6f4\" (UniqueName: \"kubernetes.io/projected/60aa603c-1807-48aa-80c0-fd56eedeff49-kube-api-access-7r6f4\") pod \"metallb-operator-webhook-server-7c64f56f84-7stp8\" (UID: \"60aa603c-1807-48aa-80c0-fd56eedeff49\") " pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:11:00 crc kubenswrapper[4894]: I1210 19:11:00.040692 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:11:00 crc kubenswrapper[4894]: I1210 19:11:00.199207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8"] Dec 10 19:11:00 crc kubenswrapper[4894]: I1210 19:11:00.282579 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8"] Dec 10 19:11:00 crc kubenswrapper[4894]: I1210 19:11:00.290145 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" event={"ID":"25e24603-77a9-43cd-86c6-941b45ab2d4b","Type":"ContainerStarted","Data":"ef67ac04d779dfb5f41d911f8f146920b753d2833bf03fa86ee16e56925223fa"} Dec 10 19:11:00 crc kubenswrapper[4894]: W1210 19:11:00.293584 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60aa603c_1807_48aa_80c0_fd56eedeff49.slice/crio-b1b89d1ee3aeefcd61c244ce7d71e4c281a756038f0214dd26f8bdafb1f4f2e1 WatchSource:0}: Error finding container b1b89d1ee3aeefcd61c244ce7d71e4c281a756038f0214dd26f8bdafb1f4f2e1: Status 404 returned error can't find the container with id b1b89d1ee3aeefcd61c244ce7d71e4c281a756038f0214dd26f8bdafb1f4f2e1 Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.298102 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" event={"ID":"60aa603c-1807-48aa-80c0-fd56eedeff49","Type":"ContainerStarted","Data":"b1b89d1ee3aeefcd61c244ce7d71e4c281a756038f0214dd26f8bdafb1f4f2e1"} Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.310032 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zhk9l"] Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.311651 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.315960 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zhk9l"] Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.392121 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-catalog-content\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.392203 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-utilities\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.392252 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk2jp\" (UniqueName: \"kubernetes.io/projected/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-kube-api-access-jk2jp\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.493890 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-utilities\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.493973 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk2jp\" (UniqueName: \"kubernetes.io/projected/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-kube-api-access-jk2jp\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.494001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-catalog-content\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.494745 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-utilities\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.494783 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-catalog-content\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.515231 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk2jp\" (UniqueName: \"kubernetes.io/projected/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-kube-api-access-jk2jp\") pod \"certified-operators-zhk9l\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.664524 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:01 crc kubenswrapper[4894]: I1210 19:11:01.981768 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zhk9l"] Dec 10 19:11:02 crc kubenswrapper[4894]: I1210 19:11:02.308137 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerID="9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898" exitCode=0 Dec 10 19:11:02 crc kubenswrapper[4894]: I1210 19:11:02.308304 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zhk9l" event={"ID":"5ed3e946-0ffb-4b0e-90eb-cac3d44693df","Type":"ContainerDied","Data":"9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898"} Dec 10 19:11:02 crc kubenswrapper[4894]: I1210 19:11:02.308514 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zhk9l" event={"ID":"5ed3e946-0ffb-4b0e-90eb-cac3d44693df","Type":"ContainerStarted","Data":"8b7239a0506f496356cc9fa7aa3cc9fb10070ab6541c4ab44fb01c25f065054a"} Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.409166 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" event={"ID":"60aa603c-1807-48aa-80c0-fd56eedeff49","Type":"ContainerStarted","Data":"493f01155cc0504f8a7e79d2a6e879ce0a851b1bed40f104c487f2476de9a810"} Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.409523 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.411761 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerID="3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0" exitCode=0 Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.411835 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zhk9l" event={"ID":"5ed3e946-0ffb-4b0e-90eb-cac3d44693df","Type":"ContainerDied","Data":"3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0"} Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.414917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" event={"ID":"25e24603-77a9-43cd-86c6-941b45ab2d4b","Type":"ContainerStarted","Data":"ef20affa4ba55f58a563889858b67fae1524cd7db6853773ed35d636fd7e160a"} Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.415308 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.428135 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" podStartSLOduration=2.189491737 podStartE2EDuration="8.428118023s" podCreationTimestamp="2025-12-10 19:10:59 +0000 UTC" firstStartedPulling="2025-12-10 19:11:00.300475101 +0000 UTC m=+943.095322893" lastFinishedPulling="2025-12-10 19:11:06.539101387 +0000 UTC m=+949.333949179" observedRunningTime="2025-12-10 19:11:07.426547914 +0000 UTC m=+950.221395726" watchObservedRunningTime="2025-12-10 19:11:07.428118023 +0000 UTC m=+950.222965825" Dec 10 19:11:07 crc kubenswrapper[4894]: I1210 19:11:07.460787 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" podStartSLOduration=2.162695139 podStartE2EDuration="8.460765992s" podCreationTimestamp="2025-12-10 19:10:59 +0000 UTC" firstStartedPulling="2025-12-10 19:11:00.236669478 +0000 UTC m=+943.031517270" lastFinishedPulling="2025-12-10 19:11:06.534740331 +0000 UTC m=+949.329588123" observedRunningTime="2025-12-10 19:11:07.457689227 +0000 UTC m=+950.252537049" watchObservedRunningTime="2025-12-10 19:11:07.460765992 +0000 UTC m=+950.255613784" Dec 10 19:11:08 crc kubenswrapper[4894]: I1210 19:11:08.424704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zhk9l" event={"ID":"5ed3e946-0ffb-4b0e-90eb-cac3d44693df","Type":"ContainerStarted","Data":"6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396"} Dec 10 19:11:08 crc kubenswrapper[4894]: I1210 19:11:08.458511 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zhk9l" podStartSLOduration=1.9382986899999999 podStartE2EDuration="7.458482121s" podCreationTimestamp="2025-12-10 19:11:01 +0000 UTC" firstStartedPulling="2025-12-10 19:11:02.313791037 +0000 UTC m=+945.108638829" lastFinishedPulling="2025-12-10 19:11:07.833974468 +0000 UTC m=+950.628822260" observedRunningTime="2025-12-10 19:11:08.451331006 +0000 UTC m=+951.246178838" watchObservedRunningTime="2025-12-10 19:11:08.458482121 +0000 UTC m=+951.253329933" Dec 10 19:11:11 crc kubenswrapper[4894]: I1210 19:11:11.664974 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:11 crc kubenswrapper[4894]: I1210 19:11:11.665463 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:11 crc kubenswrapper[4894]: I1210 19:11:11.705724 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:19 crc kubenswrapper[4894]: I1210 19:11:19.577830 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:11:19 crc kubenswrapper[4894]: I1210 19:11:19.578664 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:11:20 crc kubenswrapper[4894]: I1210 19:11:20.049921 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7c64f56f84-7stp8" Dec 10 19:11:21 crc kubenswrapper[4894]: I1210 19:11:21.707164 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:21 crc kubenswrapper[4894]: I1210 19:11:21.770430 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zhk9l"] Dec 10 19:11:22 crc kubenswrapper[4894]: I1210 19:11:22.521734 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zhk9l" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="registry-server" containerID="cri-o://6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396" gracePeriod=2 Dec 10 19:11:22 crc kubenswrapper[4894]: I1210 19:11:22.963771 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.162342 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk2jp\" (UniqueName: \"kubernetes.io/projected/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-kube-api-access-jk2jp\") pod \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.162521 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-catalog-content\") pod \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.162559 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-utilities\") pod \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\" (UID: \"5ed3e946-0ffb-4b0e-90eb-cac3d44693df\") " Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.164459 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-utilities" (OuterVolumeSpecName: "utilities") pod "5ed3e946-0ffb-4b0e-90eb-cac3d44693df" (UID: "5ed3e946-0ffb-4b0e-90eb-cac3d44693df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.168605 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-kube-api-access-jk2jp" (OuterVolumeSpecName: "kube-api-access-jk2jp") pod "5ed3e946-0ffb-4b0e-90eb-cac3d44693df" (UID: "5ed3e946-0ffb-4b0e-90eb-cac3d44693df"). InnerVolumeSpecName "kube-api-access-jk2jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.230890 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ed3e946-0ffb-4b0e-90eb-cac3d44693df" (UID: "5ed3e946-0ffb-4b0e-90eb-cac3d44693df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.263736 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.263768 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.263782 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk2jp\" (UniqueName: \"kubernetes.io/projected/5ed3e946-0ffb-4b0e-90eb-cac3d44693df-kube-api-access-jk2jp\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.529705 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerID="6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396" exitCode=0 Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.529921 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zhk9l" event={"ID":"5ed3e946-0ffb-4b0e-90eb-cac3d44693df","Type":"ContainerDied","Data":"6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396"} Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.530296 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zhk9l" event={"ID":"5ed3e946-0ffb-4b0e-90eb-cac3d44693df","Type":"ContainerDied","Data":"8b7239a0506f496356cc9fa7aa3cc9fb10070ab6541c4ab44fb01c25f065054a"} Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.530420 4894 scope.go:117] "RemoveContainer" containerID="6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.530025 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zhk9l" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.553727 4894 scope.go:117] "RemoveContainer" containerID="3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.593791 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zhk9l"] Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.594028 4894 scope.go:117] "RemoveContainer" containerID="9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.598401 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zhk9l"] Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.623694 4894 scope.go:117] "RemoveContainer" containerID="6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396" Dec 10 19:11:23 crc kubenswrapper[4894]: E1210 19:11:23.624227 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396\": container with ID starting with 6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396 not found: ID does not exist" containerID="6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.624387 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396"} err="failed to get container status \"6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396\": rpc error: code = NotFound desc = could not find container \"6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396\": container with ID starting with 6eb0f627f4cedf6374f0f5988b2dc111dd19484b8c53f41d65a7c0118fd01396 not found: ID does not exist" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.624496 4894 scope.go:117] "RemoveContainer" containerID="3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0" Dec 10 19:11:23 crc kubenswrapper[4894]: E1210 19:11:23.624931 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0\": container with ID starting with 3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0 not found: ID does not exist" containerID="3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.625070 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0"} err="failed to get container status \"3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0\": rpc error: code = NotFound desc = could not find container \"3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0\": container with ID starting with 3383c2565384243e8aa7e1e44b899d10c6631547de38c45a121cdff5b85a4fd0 not found: ID does not exist" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.625167 4894 scope.go:117] "RemoveContainer" containerID="9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898" Dec 10 19:11:23 crc kubenswrapper[4894]: E1210 19:11:23.627328 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898\": container with ID starting with 9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898 not found: ID does not exist" containerID="9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898" Dec 10 19:11:23 crc kubenswrapper[4894]: I1210 19:11:23.627469 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898"} err="failed to get container status \"9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898\": rpc error: code = NotFound desc = could not find container \"9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898\": container with ID starting with 9de4551b61984a190b2ee37e4c3d92e2bfaa5bba51e234785e0a83f01aa4f898 not found: ID does not exist" Dec 10 19:11:25 crc kubenswrapper[4894]: I1210 19:11:25.553594 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" path="/var/lib/kubelet/pods/5ed3e946-0ffb-4b0e-90eb-cac3d44693df/volumes" Dec 10 19:11:39 crc kubenswrapper[4894]: I1210 19:11:39.790527 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8484c557fc-rnmp8" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.561433 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-svgvf"] Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.561811 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="extract-content" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.561836 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="extract-content" Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.561889 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="extract-utilities" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.561899 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="extract-utilities" Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.561915 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="registry-server" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.561923 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="registry-server" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.562062 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ed3e946-0ffb-4b0e-90eb-cac3d44693df" containerName="registry-server" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.564735 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.566439 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g"] Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.567331 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.567624 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-bq44k" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.573255 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.573428 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.573549 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.583826 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g"] Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-metrics\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631300 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghgh6\" (UniqueName: \"kubernetes.io/projected/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-kube-api-access-ghgh6\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631339 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njkcr\" (UniqueName: \"kubernetes.io/projected/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-kube-api-access-njkcr\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631368 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631408 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-metrics-certs\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631426 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-conf\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-reloader\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631472 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-startup\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.631492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-sockets\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.632416 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-fpgb2"] Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.633546 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.639636 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-stkcg" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.639868 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.640004 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.640118 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.665700 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-q2tvc"] Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.666947 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.668840 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.677081 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-q2tvc"] Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.732653 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-metrics\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.732700 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxvhn\" (UniqueName: \"kubernetes.io/projected/eb70538a-122e-480f-8853-15e96afdf9a4-kube-api-access-lxvhn\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.732725 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlk8f\" (UniqueName: \"kubernetes.io/projected/ffb10c89-1b21-48b2-8d11-320ccf17dd64-kube-api-access-xlk8f\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.732755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghgh6\" (UniqueName: \"kubernetes.io/projected/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-kube-api-access-ghgh6\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.732827 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njkcr\" (UniqueName: \"kubernetes.io/projected/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-kube-api-access-njkcr\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-metrics-certs\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733090 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733117 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70538a-122e-480f-8853-15e96afdf9a4-cert\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733141 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ffb10c89-1b21-48b2-8d11-320ccf17dd64-metallb-excludel2\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733160 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-metrics-certs\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733178 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-conf\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733201 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-reloader\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.733203 4894 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733220 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-startup\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.733257 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-cert podName:a3852d90-39db-4a6e-b4cd-4d805d18a2b5 nodeName:}" failed. No retries permitted until 2025-12-10 19:11:41.233242621 +0000 UTC m=+984.028090413 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-cert") pod "frr-k8s-webhook-server-7fcb986d4-skk4g" (UID: "a3852d90-39db-4a6e-b4cd-4d805d18a2b5") : secret "frr-k8s-webhook-server-cert" not found Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733297 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-sockets\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733328 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb70538a-122e-480f-8853-15e96afdf9a4-metrics-certs\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-conf\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.733822 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-sockets\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.734532 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-frr-startup\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.734762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-metrics\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.734910 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-reloader\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.750441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghgh6\" (UniqueName: \"kubernetes.io/projected/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-kube-api-access-ghgh6\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.754703 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njkcr\" (UniqueName: \"kubernetes.io/projected/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-kube-api-access-njkcr\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.756914 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee-metrics-certs\") pod \"frr-k8s-svgvf\" (UID: \"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee\") " pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834236 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834284 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb70538a-122e-480f-8853-15e96afdf9a4-metrics-certs\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834316 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxvhn\" (UniqueName: \"kubernetes.io/projected/eb70538a-122e-480f-8853-15e96afdf9a4-kube-api-access-lxvhn\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlk8f\" (UniqueName: \"kubernetes.io/projected/ffb10c89-1b21-48b2-8d11-320ccf17dd64-kube-api-access-xlk8f\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834378 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-metrics-certs\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834415 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70538a-122e-480f-8853-15e96afdf9a4-cert\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.834432 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ffb10c89-1b21-48b2-8d11-320ccf17dd64-metallb-excludel2\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.834463 4894 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 19:11:40 crc kubenswrapper[4894]: E1210 19:11:40.834550 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist podName:ffb10c89-1b21-48b2-8d11-320ccf17dd64 nodeName:}" failed. No retries permitted until 2025-12-10 19:11:41.334528045 +0000 UTC m=+984.129375837 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist") pod "speaker-fpgb2" (UID: "ffb10c89-1b21-48b2-8d11-320ccf17dd64") : secret "metallb-memberlist" not found Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.835089 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ffb10c89-1b21-48b2-8d11-320ccf17dd64-metallb-excludel2\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.837401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb70538a-122e-480f-8853-15e96afdf9a4-metrics-certs\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.839598 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-metrics-certs\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.843187 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.847957 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70538a-122e-480f-8853-15e96afdf9a4-cert\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.851589 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlk8f\" (UniqueName: \"kubernetes.io/projected/ffb10c89-1b21-48b2-8d11-320ccf17dd64-kube-api-access-xlk8f\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.854266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxvhn\" (UniqueName: \"kubernetes.io/projected/eb70538a-122e-480f-8853-15e96afdf9a4-kube-api-access-lxvhn\") pod \"controller-f8648f98b-q2tvc\" (UID: \"eb70538a-122e-480f-8853-15e96afdf9a4\") " pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.895274 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:40 crc kubenswrapper[4894]: I1210 19:11:40.999466 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.238659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.244585 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a3852d90-39db-4a6e-b4cd-4d805d18a2b5-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-skk4g\" (UID: \"a3852d90-39db-4a6e-b4cd-4d805d18a2b5\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.340783 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:41 crc kubenswrapper[4894]: E1210 19:11:41.341012 4894 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 19:11:41 crc kubenswrapper[4894]: E1210 19:11:41.341113 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist podName:ffb10c89-1b21-48b2-8d11-320ccf17dd64 nodeName:}" failed. No retries permitted until 2025-12-10 19:11:42.341091914 +0000 UTC m=+985.135939716 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist") pod "speaker-fpgb2" (UID: "ffb10c89-1b21-48b2-8d11-320ccf17dd64") : secret "metallb-memberlist" not found Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.506359 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.529969 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-q2tvc"] Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.668540 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"a2fa4fbcdc9699d600043c4d4b281abd9736fe697ff4483ed0e39fb061ce4298"} Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.669930 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-q2tvc" event={"ID":"eb70538a-122e-480f-8853-15e96afdf9a4","Type":"ContainerStarted","Data":"c88c9912195f1b6f5c965a87ec7c8a53793b8815a40ccd4077b60f78e5cf0530"} Dec 10 19:11:41 crc kubenswrapper[4894]: I1210 19:11:41.746354 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g"] Dec 10 19:11:41 crc kubenswrapper[4894]: W1210 19:11:41.769259 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3852d90_39db_4a6e_b4cd_4d805d18a2b5.slice/crio-11910c8f4bbdbede83351c005f10ce8963b4f393d4ddf1a720676d53be7f529d WatchSource:0}: Error finding container 11910c8f4bbdbede83351c005f10ce8963b4f393d4ddf1a720676d53be7f529d: Status 404 returned error can't find the container with id 11910c8f4bbdbede83351c005f10ce8963b4f393d4ddf1a720676d53be7f529d Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.357447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.366394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ffb10c89-1b21-48b2-8d11-320ccf17dd64-memberlist\") pod \"speaker-fpgb2\" (UID: \"ffb10c89-1b21-48b2-8d11-320ccf17dd64\") " pod="metallb-system/speaker-fpgb2" Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.454641 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fpgb2" Dec 10 19:11:42 crc kubenswrapper[4894]: W1210 19:11:42.486596 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffb10c89_1b21_48b2_8d11_320ccf17dd64.slice/crio-390e1ccc826165b381f3cf3d1a5197e94983dbf08ead6d4e1a02c23785a74222 WatchSource:0}: Error finding container 390e1ccc826165b381f3cf3d1a5197e94983dbf08ead6d4e1a02c23785a74222: Status 404 returned error can't find the container with id 390e1ccc826165b381f3cf3d1a5197e94983dbf08ead6d4e1a02c23785a74222 Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.677139 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fpgb2" event={"ID":"ffb10c89-1b21-48b2-8d11-320ccf17dd64","Type":"ContainerStarted","Data":"390e1ccc826165b381f3cf3d1a5197e94983dbf08ead6d4e1a02c23785a74222"} Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.679277 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-q2tvc" event={"ID":"eb70538a-122e-480f-8853-15e96afdf9a4","Type":"ContainerStarted","Data":"d0badb8a95a4778afd8608a8a9126cd0c6309719ccc2f4f107f135e3a0754eb8"} Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.679305 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-q2tvc" event={"ID":"eb70538a-122e-480f-8853-15e96afdf9a4","Type":"ContainerStarted","Data":"9b22e08effb914de1df3f9250d13b976ef99a8d359b3dafbdfc7a89853c40a11"} Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.679391 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.682710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" event={"ID":"a3852d90-39db-4a6e-b4cd-4d805d18a2b5","Type":"ContainerStarted","Data":"11910c8f4bbdbede83351c005f10ce8963b4f393d4ddf1a720676d53be7f529d"} Dec 10 19:11:42 crc kubenswrapper[4894]: I1210 19:11:42.705711 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-q2tvc" podStartSLOduration=2.705690359 podStartE2EDuration="2.705690359s" podCreationTimestamp="2025-12-10 19:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:11:42.701499125 +0000 UTC m=+985.496346907" watchObservedRunningTime="2025-12-10 19:11:42.705690359 +0000 UTC m=+985.500538151" Dec 10 19:11:43 crc kubenswrapper[4894]: I1210 19:11:43.695250 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fpgb2" event={"ID":"ffb10c89-1b21-48b2-8d11-320ccf17dd64","Type":"ContainerStarted","Data":"d1a8eea94cb7d4be198a31e844cd6b06b6dcb3d2aa24d62b477bc516e7955bd8"} Dec 10 19:11:43 crc kubenswrapper[4894]: I1210 19:11:43.695565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fpgb2" event={"ID":"ffb10c89-1b21-48b2-8d11-320ccf17dd64","Type":"ContainerStarted","Data":"fb7cd30b7911facaa77b2b9750348ea7be4a133306045adcf8e9017cd71b6942"} Dec 10 19:11:43 crc kubenswrapper[4894]: I1210 19:11:43.695616 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-fpgb2" Dec 10 19:11:43 crc kubenswrapper[4894]: I1210 19:11:43.718278 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-fpgb2" podStartSLOduration=3.718259911 podStartE2EDuration="3.718259911s" podCreationTimestamp="2025-12-10 19:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:11:43.717584392 +0000 UTC m=+986.512432184" watchObservedRunningTime="2025-12-10 19:11:43.718259911 +0000 UTC m=+986.513107703" Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.576985 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.577551 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.577600 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.578187 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0a333f57d00f6ace0bfa8b35a6220c82d5630155e1eac3a22921478ab8cbce3"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.578251 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://a0a333f57d00f6ace0bfa8b35a6220c82d5630155e1eac3a22921478ab8cbce3" gracePeriod=600 Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.745443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" event={"ID":"a3852d90-39db-4a6e-b4cd-4d805d18a2b5","Type":"ContainerStarted","Data":"91c1f4678ce35defd6ce34d7dfa1ff17e38d9ec6a48944b96460761ea1ff65f6"} Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.745748 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.747617 4894 generic.go:334] "Generic (PLEG): container finished" podID="3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee" containerID="b0c85a304d7b1b6dafc27a217f4233ded6ec83e8265457a34ac29ca1d2bdbad1" exitCode=0 Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.747647 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerDied","Data":"b0c85a304d7b1b6dafc27a217f4233ded6ec83e8265457a34ac29ca1d2bdbad1"} Dec 10 19:11:49 crc kubenswrapper[4894]: I1210 19:11:49.781127 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" podStartSLOduration=2.719804443 podStartE2EDuration="9.781103806s" podCreationTimestamp="2025-12-10 19:11:40 +0000 UTC" firstStartedPulling="2025-12-10 19:11:41.771610463 +0000 UTC m=+984.566458245" lastFinishedPulling="2025-12-10 19:11:48.832909816 +0000 UTC m=+991.627757608" observedRunningTime="2025-12-10 19:11:49.760570897 +0000 UTC m=+992.555418699" watchObservedRunningTime="2025-12-10 19:11:49.781103806 +0000 UTC m=+992.575951628" Dec 10 19:11:50 crc kubenswrapper[4894]: I1210 19:11:50.758180 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="a0a333f57d00f6ace0bfa8b35a6220c82d5630155e1eac3a22921478ab8cbce3" exitCode=0 Dec 10 19:11:50 crc kubenswrapper[4894]: I1210 19:11:50.758251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"a0a333f57d00f6ace0bfa8b35a6220c82d5630155e1eac3a22921478ab8cbce3"} Dec 10 19:11:50 crc kubenswrapper[4894]: I1210 19:11:50.758719 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"74b5960f63532f14bffacaae454d958ec29335c9f8adf904077f837019704bae"} Dec 10 19:11:50 crc kubenswrapper[4894]: I1210 19:11:50.758757 4894 scope.go:117] "RemoveContainer" containerID="4ec2353aa9d8d229a346f66ddc339bc5b86fce1f5b3bf7d51741a93c0c96e096" Dec 10 19:11:50 crc kubenswrapper[4894]: I1210 19:11:50.762761 4894 generic.go:334] "Generic (PLEG): container finished" podID="3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee" containerID="c14f9be1f5494c79c6786419984a38c35af232bbe6505598cdf752c5fd28677c" exitCode=0 Dec 10 19:11:50 crc kubenswrapper[4894]: I1210 19:11:50.762883 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerDied","Data":"c14f9be1f5494c79c6786419984a38c35af232bbe6505598cdf752c5fd28677c"} Dec 10 19:11:51 crc kubenswrapper[4894]: I1210 19:11:51.772428 4894 generic.go:334] "Generic (PLEG): container finished" podID="3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee" containerID="459a2c8cc71ed227a7785f039e4712ea680366dae1fa78167d67e95e663636a6" exitCode=0 Dec 10 19:11:51 crc kubenswrapper[4894]: I1210 19:11:51.772505 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerDied","Data":"459a2c8cc71ed227a7785f039e4712ea680366dae1fa78167d67e95e663636a6"} Dec 10 19:11:52 crc kubenswrapper[4894]: I1210 19:11:52.458375 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-fpgb2" Dec 10 19:11:52 crc kubenswrapper[4894]: I1210 19:11:52.789152 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"a15de1acb21ffc14a7fd6eebc301a078abc298d2502f27c31fe3d91ce767de39"} Dec 10 19:11:52 crc kubenswrapper[4894]: I1210 19:11:52.789198 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"ec5dec31b5c610b5ff52dcd440c67fc2c6d4c45b7708c5a87567119f875185ec"} Dec 10 19:11:52 crc kubenswrapper[4894]: I1210 19:11:52.789210 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"840e7745bea3a42916566ec500717fdefbea33393950cc484ec94d788a59d569"} Dec 10 19:11:52 crc kubenswrapper[4894]: I1210 19:11:52.789220 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"79cc2d3974512b425748abe59789f398f24584688fcf1c9dd5d4ebedc427101c"} Dec 10 19:11:52 crc kubenswrapper[4894]: I1210 19:11:52.789230 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"3b73699912c17c19e6cedd8aae624e43b0528e1627b3c0a36044d43981484e1e"} Dec 10 19:11:53 crc kubenswrapper[4894]: I1210 19:11:53.803114 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-svgvf" event={"ID":"3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee","Type":"ContainerStarted","Data":"10f793e8fed5febf7e0ec862153edd24a40fccdd40c5772764ce899db3904a95"} Dec 10 19:11:53 crc kubenswrapper[4894]: I1210 19:11:53.803872 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:53 crc kubenswrapper[4894]: I1210 19:11:53.837655 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-svgvf" podStartSLOduration=6.051875033 podStartE2EDuration="13.83762959s" podCreationTimestamp="2025-12-10 19:11:40 +0000 UTC" firstStartedPulling="2025-12-10 19:11:41.029526959 +0000 UTC m=+983.824374751" lastFinishedPulling="2025-12-10 19:11:48.815281496 +0000 UTC m=+991.610129308" observedRunningTime="2025-12-10 19:11:53.828947574 +0000 UTC m=+996.623795386" watchObservedRunningTime="2025-12-10 19:11:53.83762959 +0000 UTC m=+996.632477422" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.332929 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wtm98"] Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.334754 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.338508 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-27hzb" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.338523 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.338598 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.343656 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wtm98"] Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.360056 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7blcb\" (UniqueName: \"kubernetes.io/projected/5d693e7f-3281-4738-8232-b2a3bfc727a9-kube-api-access-7blcb\") pod \"openstack-operator-index-wtm98\" (UID: \"5d693e7f-3281-4738-8232-b2a3bfc727a9\") " pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.461762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7blcb\" (UniqueName: \"kubernetes.io/projected/5d693e7f-3281-4738-8232-b2a3bfc727a9-kube-api-access-7blcb\") pod \"openstack-operator-index-wtm98\" (UID: \"5d693e7f-3281-4738-8232-b2a3bfc727a9\") " pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.487958 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7blcb\" (UniqueName: \"kubernetes.io/projected/5d693e7f-3281-4738-8232-b2a3bfc727a9-kube-api-access-7blcb\") pod \"openstack-operator-index-wtm98\" (UID: \"5d693e7f-3281-4738-8232-b2a3bfc727a9\") " pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.661020 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.896239 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:55 crc kubenswrapper[4894]: I1210 19:11:55.931155 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-svgvf" Dec 10 19:11:56 crc kubenswrapper[4894]: I1210 19:11:56.077235 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wtm98"] Dec 10 19:11:56 crc kubenswrapper[4894]: I1210 19:11:56.840970 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wtm98" event={"ID":"5d693e7f-3281-4738-8232-b2a3bfc727a9","Type":"ContainerStarted","Data":"87ada7dc98e6f2b89f624d71d89b371187163c5d2ed6c1dc78ad4b7be60f7ff9"} Dec 10 19:11:58 crc kubenswrapper[4894]: I1210 19:11:58.508475 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wtm98"] Dec 10 19:11:58 crc kubenswrapper[4894]: I1210 19:11:58.862654 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-wtm98" podUID="5d693e7f-3281-4738-8232-b2a3bfc727a9" containerName="registry-server" containerID="cri-o://0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0" gracePeriod=2 Dec 10 19:11:58 crc kubenswrapper[4894]: I1210 19:11:58.862452 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wtm98" event={"ID":"5d693e7f-3281-4738-8232-b2a3bfc727a9","Type":"ContainerStarted","Data":"0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0"} Dec 10 19:11:58 crc kubenswrapper[4894]: I1210 19:11:58.887952 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wtm98" podStartSLOduration=1.351739183 podStartE2EDuration="3.887926745s" podCreationTimestamp="2025-12-10 19:11:55 +0000 UTC" firstStartedPulling="2025-12-10 19:11:56.085692766 +0000 UTC m=+998.880540558" lastFinishedPulling="2025-12-10 19:11:58.621880328 +0000 UTC m=+1001.416728120" observedRunningTime="2025-12-10 19:11:58.879519476 +0000 UTC m=+1001.674367308" watchObservedRunningTime="2025-12-10 19:11:58.887926745 +0000 UTC m=+1001.682774607" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.116598 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jz862"] Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.117363 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.133024 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jz862"] Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.246075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dml4k\" (UniqueName: \"kubernetes.io/projected/3a2552c3-74d8-4da0-aae9-12bb7b08a42b-kube-api-access-dml4k\") pod \"openstack-operator-index-jz862\" (UID: \"3a2552c3-74d8-4da0-aae9-12bb7b08a42b\") " pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.324229 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.347600 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dml4k\" (UniqueName: \"kubernetes.io/projected/3a2552c3-74d8-4da0-aae9-12bb7b08a42b-kube-api-access-dml4k\") pod \"openstack-operator-index-jz862\" (UID: \"3a2552c3-74d8-4da0-aae9-12bb7b08a42b\") " pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.370100 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dml4k\" (UniqueName: \"kubernetes.io/projected/3a2552c3-74d8-4da0-aae9-12bb7b08a42b-kube-api-access-dml4k\") pod \"openstack-operator-index-jz862\" (UID: \"3a2552c3-74d8-4da0-aae9-12bb7b08a42b\") " pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.448552 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7blcb\" (UniqueName: \"kubernetes.io/projected/5d693e7f-3281-4738-8232-b2a3bfc727a9-kube-api-access-7blcb\") pod \"5d693e7f-3281-4738-8232-b2a3bfc727a9\" (UID: \"5d693e7f-3281-4738-8232-b2a3bfc727a9\") " Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.452759 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.456178 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d693e7f-3281-4738-8232-b2a3bfc727a9-kube-api-access-7blcb" (OuterVolumeSpecName: "kube-api-access-7blcb") pod "5d693e7f-3281-4738-8232-b2a3bfc727a9" (UID: "5d693e7f-3281-4738-8232-b2a3bfc727a9"). InnerVolumeSpecName "kube-api-access-7blcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.550943 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7blcb\" (UniqueName: \"kubernetes.io/projected/5d693e7f-3281-4738-8232-b2a3bfc727a9-kube-api-access-7blcb\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.863622 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jz862"] Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.872045 4894 generic.go:334] "Generic (PLEG): container finished" podID="5d693e7f-3281-4738-8232-b2a3bfc727a9" containerID="0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0" exitCode=0 Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.872114 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wtm98" event={"ID":"5d693e7f-3281-4738-8232-b2a3bfc727a9","Type":"ContainerDied","Data":"0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0"} Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.872173 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wtm98" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.872205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wtm98" event={"ID":"5d693e7f-3281-4738-8232-b2a3bfc727a9","Type":"ContainerDied","Data":"87ada7dc98e6f2b89f624d71d89b371187163c5d2ed6c1dc78ad4b7be60f7ff9"} Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.872231 4894 scope.go:117] "RemoveContainer" containerID="0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.914572 4894 scope.go:117] "RemoveContainer" containerID="0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0" Dec 10 19:11:59 crc kubenswrapper[4894]: E1210 19:11:59.915634 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0\": container with ID starting with 0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0 not found: ID does not exist" containerID="0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.915671 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0"} err="failed to get container status \"0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0\": rpc error: code = NotFound desc = could not find container \"0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0\": container with ID starting with 0e59c21d4c20b6a028203ad88707d3f6ac351fe032ec8b886ebd43d9c9abc2d0 not found: ID does not exist" Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.918189 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wtm98"] Dec 10 19:11:59 crc kubenswrapper[4894]: I1210 19:11:59.922023 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-wtm98"] Dec 10 19:12:00 crc kubenswrapper[4894]: I1210 19:12:00.884246 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jz862" event={"ID":"3a2552c3-74d8-4da0-aae9-12bb7b08a42b","Type":"ContainerStarted","Data":"44ff0ebd1ef78aa7ffd5749224ee5d2854bd2b2386c24dcd388f9ea4f0b40a32"} Dec 10 19:12:00 crc kubenswrapper[4894]: I1210 19:12:00.884290 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jz862" event={"ID":"3a2552c3-74d8-4da0-aae9-12bb7b08a42b","Type":"ContainerStarted","Data":"b8c09e8839ad13870bdf1fd91bd8ecb469550039fface2ce555ae66188d5924c"} Dec 10 19:12:00 crc kubenswrapper[4894]: I1210 19:12:00.902148 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jz862" podStartSLOduration=1.833952355 podStartE2EDuration="1.902102399s" podCreationTimestamp="2025-12-10 19:11:59 +0000 UTC" firstStartedPulling="2025-12-10 19:11:59.869540944 +0000 UTC m=+1002.664388776" lastFinishedPulling="2025-12-10 19:11:59.937691028 +0000 UTC m=+1002.732538820" observedRunningTime="2025-12-10 19:12:00.901584605 +0000 UTC m=+1003.696432427" watchObservedRunningTime="2025-12-10 19:12:00.902102399 +0000 UTC m=+1003.696950201" Dec 10 19:12:01 crc kubenswrapper[4894]: I1210 19:12:01.002104 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-q2tvc" Dec 10 19:12:01 crc kubenswrapper[4894]: I1210 19:12:01.513818 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-skk4g" Dec 10 19:12:01 crc kubenswrapper[4894]: I1210 19:12:01.560247 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d693e7f-3281-4738-8232-b2a3bfc727a9" path="/var/lib/kubelet/pods/5d693e7f-3281-4738-8232-b2a3bfc727a9/volumes" Dec 10 19:12:09 crc kubenswrapper[4894]: I1210 19:12:09.453502 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:12:09 crc kubenswrapper[4894]: I1210 19:12:09.454186 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:12:09 crc kubenswrapper[4894]: I1210 19:12:09.500619 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:12:09 crc kubenswrapper[4894]: I1210 19:12:09.993668 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-jz862" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.761888 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x"] Dec 10 19:12:10 crc kubenswrapper[4894]: E1210 19:12:10.762179 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d693e7f-3281-4738-8232-b2a3bfc727a9" containerName="registry-server" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.762193 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d693e7f-3281-4738-8232-b2a3bfc727a9" containerName="registry-server" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.762357 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d693e7f-3281-4738-8232-b2a3bfc727a9" containerName="registry-server" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.763441 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.768578 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qfkmv" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.795151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x"] Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.899968 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-svgvf" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.924332 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-util\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.924414 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-bundle\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:10 crc kubenswrapper[4894]: I1210 19:12:10.924479 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftxx2\" (UniqueName: \"kubernetes.io/projected/e180e484-98a3-4635-9b27-fdaa1687d3f6-kube-api-access-ftxx2\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.026586 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-util\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.026694 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-bundle\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.026746 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftxx2\" (UniqueName: \"kubernetes.io/projected/e180e484-98a3-4635-9b27-fdaa1687d3f6-kube-api-access-ftxx2\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.027277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-util\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.027327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-bundle\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.054279 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftxx2\" (UniqueName: \"kubernetes.io/projected/e180e484-98a3-4635-9b27-fdaa1687d3f6-kube-api-access-ftxx2\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.090765 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.411913 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x"] Dec 10 19:12:11 crc kubenswrapper[4894]: W1210 19:12:11.418076 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode180e484_98a3_4635_9b27_fdaa1687d3f6.slice/crio-3fc0c814929d8feee40e34ba5be9283494b84448de465f4300cd3a52b3e257f6 WatchSource:0}: Error finding container 3fc0c814929d8feee40e34ba5be9283494b84448de465f4300cd3a52b3e257f6: Status 404 returned error can't find the container with id 3fc0c814929d8feee40e34ba5be9283494b84448de465f4300cd3a52b3e257f6 Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.984960 4894 generic.go:334] "Generic (PLEG): container finished" podID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerID="08aff06df294b33c7560cf569811c84c879ab7b2e8fc12e95410e3b2af987c0e" exitCode=0 Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.985000 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" event={"ID":"e180e484-98a3-4635-9b27-fdaa1687d3f6","Type":"ContainerDied","Data":"08aff06df294b33c7560cf569811c84c879ab7b2e8fc12e95410e3b2af987c0e"} Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.985024 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" event={"ID":"e180e484-98a3-4635-9b27-fdaa1687d3f6","Type":"ContainerStarted","Data":"3fc0c814929d8feee40e34ba5be9283494b84448de465f4300cd3a52b3e257f6"} Dec 10 19:12:11 crc kubenswrapper[4894]: I1210 19:12:11.987877 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:12:12 crc kubenswrapper[4894]: I1210 19:12:12.997429 4894 generic.go:334] "Generic (PLEG): container finished" podID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerID="d6abd034c2d1b096f064ea8431e217a8f9b697dfe40ada1f5db320c6c448f146" exitCode=0 Dec 10 19:12:12 crc kubenswrapper[4894]: I1210 19:12:12.997554 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" event={"ID":"e180e484-98a3-4635-9b27-fdaa1687d3f6","Type":"ContainerDied","Data":"d6abd034c2d1b096f064ea8431e217a8f9b697dfe40ada1f5db320c6c448f146"} Dec 10 19:12:14 crc kubenswrapper[4894]: I1210 19:12:14.010506 4894 generic.go:334] "Generic (PLEG): container finished" podID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerID="0895de9a8a597a7a885e7624813985eb1731e3d84bf012b31f8d4f2a816a7723" exitCode=0 Dec 10 19:12:14 crc kubenswrapper[4894]: I1210 19:12:14.010575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" event={"ID":"e180e484-98a3-4635-9b27-fdaa1687d3f6","Type":"ContainerDied","Data":"0895de9a8a597a7a885e7624813985eb1731e3d84bf012b31f8d4f2a816a7723"} Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.341087 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.492072 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftxx2\" (UniqueName: \"kubernetes.io/projected/e180e484-98a3-4635-9b27-fdaa1687d3f6-kube-api-access-ftxx2\") pod \"e180e484-98a3-4635-9b27-fdaa1687d3f6\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.492211 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-util\") pod \"e180e484-98a3-4635-9b27-fdaa1687d3f6\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.492403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-bundle\") pod \"e180e484-98a3-4635-9b27-fdaa1687d3f6\" (UID: \"e180e484-98a3-4635-9b27-fdaa1687d3f6\") " Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.493523 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-bundle" (OuterVolumeSpecName: "bundle") pod "e180e484-98a3-4635-9b27-fdaa1687d3f6" (UID: "e180e484-98a3-4635-9b27-fdaa1687d3f6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.504595 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e180e484-98a3-4635-9b27-fdaa1687d3f6-kube-api-access-ftxx2" (OuterVolumeSpecName: "kube-api-access-ftxx2") pod "e180e484-98a3-4635-9b27-fdaa1687d3f6" (UID: "e180e484-98a3-4635-9b27-fdaa1687d3f6"). InnerVolumeSpecName "kube-api-access-ftxx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.528745 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-util" (OuterVolumeSpecName: "util") pod "e180e484-98a3-4635-9b27-fdaa1687d3f6" (UID: "e180e484-98a3-4635-9b27-fdaa1687d3f6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.594138 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.594192 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftxx2\" (UniqueName: \"kubernetes.io/projected/e180e484-98a3-4635-9b27-fdaa1687d3f6-kube-api-access-ftxx2\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:15 crc kubenswrapper[4894]: I1210 19:12:15.594214 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e180e484-98a3-4635-9b27-fdaa1687d3f6-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:16 crc kubenswrapper[4894]: I1210 19:12:16.036916 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" event={"ID":"e180e484-98a3-4635-9b27-fdaa1687d3f6","Type":"ContainerDied","Data":"3fc0c814929d8feee40e34ba5be9283494b84448de465f4300cd3a52b3e257f6"} Dec 10 19:12:16 crc kubenswrapper[4894]: I1210 19:12:16.036992 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fc0c814929d8feee40e34ba5be9283494b84448de465f4300cd3a52b3e257f6" Dec 10 19:12:16 crc kubenswrapper[4894]: I1210 19:12:16.036991 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.027827 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr"] Dec 10 19:12:23 crc kubenswrapper[4894]: E1210 19:12:23.028579 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="util" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.028592 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="util" Dec 10 19:12:23 crc kubenswrapper[4894]: E1210 19:12:23.028607 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="pull" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.028612 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="pull" Dec 10 19:12:23 crc kubenswrapper[4894]: E1210 19:12:23.028619 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="extract" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.028625 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="extract" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.028740 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e180e484-98a3-4635-9b27-fdaa1687d3f6" containerName="extract" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.029185 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.030776 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-ttv8m" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.063711 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr"] Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.111595 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2hlw\" (UniqueName: \"kubernetes.io/projected/1b3de11c-140a-427c-a958-a534599df8d9-kube-api-access-r2hlw\") pod \"openstack-operator-controller-operator-6b7b77566b-tszmr\" (UID: \"1b3de11c-140a-427c-a958-a534599df8d9\") " pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.213306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2hlw\" (UniqueName: \"kubernetes.io/projected/1b3de11c-140a-427c-a958-a534599df8d9-kube-api-access-r2hlw\") pod \"openstack-operator-controller-operator-6b7b77566b-tszmr\" (UID: \"1b3de11c-140a-427c-a958-a534599df8d9\") " pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.240015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2hlw\" (UniqueName: \"kubernetes.io/projected/1b3de11c-140a-427c-a958-a534599df8d9-kube-api-access-r2hlw\") pod \"openstack-operator-controller-operator-6b7b77566b-tszmr\" (UID: \"1b3de11c-140a-427c-a958-a534599df8d9\") " pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.347434 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:23 crc kubenswrapper[4894]: I1210 19:12:23.825354 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr"] Dec 10 19:12:24 crc kubenswrapper[4894]: I1210 19:12:24.103917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" event={"ID":"1b3de11c-140a-427c-a958-a534599df8d9","Type":"ContainerStarted","Data":"cbe9a621acabf085815987f979e1346f973dd35dbe90226ae4fb3a1023b7b151"} Dec 10 19:12:29 crc kubenswrapper[4894]: I1210 19:12:29.143364 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" event={"ID":"1b3de11c-140a-427c-a958-a534599df8d9","Type":"ContainerStarted","Data":"ef32c7cbabe4ce36b2d566a90ef9a8d290a75cdd69319e64dc659f300b9b3f7a"} Dec 10 19:12:29 crc kubenswrapper[4894]: I1210 19:12:29.144051 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:29 crc kubenswrapper[4894]: I1210 19:12:29.204576 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" podStartSLOduration=1.266534126 podStartE2EDuration="6.204553626s" podCreationTimestamp="2025-12-10 19:12:23 +0000 UTC" firstStartedPulling="2025-12-10 19:12:23.835344329 +0000 UTC m=+1026.630192161" lastFinishedPulling="2025-12-10 19:12:28.773363829 +0000 UTC m=+1031.568211661" observedRunningTime="2025-12-10 19:12:29.194553944 +0000 UTC m=+1031.989401747" watchObservedRunningTime="2025-12-10 19:12:29.204553626 +0000 UTC m=+1031.999401428" Dec 10 19:12:33 crc kubenswrapper[4894]: I1210 19:12:33.350246 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-tszmr" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.495421 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.497487 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.500078 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hplwh" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.503964 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.504954 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.506431 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-zw9h5" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.507994 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.529729 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.532756 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.537204 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-mm544" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.539771 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.546011 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.557737 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.562382 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-v9l82" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.562555 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.574086 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.589910 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.590916 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.596108 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-d2wvs" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.599568 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.623546 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.624800 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.632287 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-9cjb7" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.633298 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tnbx\" (UniqueName: \"kubernetes.io/projected/2fd92e49-8ea8-44c1-b516-a47c2c9bc516-kube-api-access-6tnbx\") pod \"cinder-operator-controller-manager-6c677c69b-kkcfl\" (UID: \"2fd92e49-8ea8-44c1-b516-a47c2c9bc516\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.633367 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2l2j\" (UniqueName: \"kubernetes.io/projected/6a960bb0-9f2c-4610-a418-791038fd1a06-kube-api-access-d2l2j\") pod \"glance-operator-controller-manager-5697bb5779-xb72b\" (UID: \"6a960bb0-9f2c-4610-a418-791038fd1a06\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.633400 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhzcm\" (UniqueName: \"kubernetes.io/projected/badfd685-2701-4ced-a77e-6eb7b7ff8c01-kube-api-access-qhzcm\") pod \"designate-operator-controller-manager-697fb699cf-c7qtj\" (UID: \"badfd685-2701-4ced-a77e-6eb7b7ff8c01\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.633466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fsqq\" (UniqueName: \"kubernetes.io/projected/1a65ee6c-f87e-4183-8c8c-9582347febda-kube-api-access-6fsqq\") pod \"heat-operator-controller-manager-5f64f6f8bb-lq68m\" (UID: \"1a65ee6c-f87e-4183-8c8c-9582347febda\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.633488 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zwb6\" (UniqueName: \"kubernetes.io/projected/3a81ced4-fe81-41ca-9623-243b811b3491-kube-api-access-4zwb6\") pod \"barbican-operator-controller-manager-7d9dfd778-99gdz\" (UID: \"3a81ced4-fe81-41ca-9623-243b811b3491\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.670529 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.684444 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.710224 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.720129 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.726972 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.728561 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.730026 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.731655 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vx7jp" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.734352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fsqq\" (UniqueName: \"kubernetes.io/projected/1a65ee6c-f87e-4183-8c8c-9582347febda-kube-api-access-6fsqq\") pod \"heat-operator-controller-manager-5f64f6f8bb-lq68m\" (UID: \"1a65ee6c-f87e-4183-8c8c-9582347febda\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.734389 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zwb6\" (UniqueName: \"kubernetes.io/projected/3a81ced4-fe81-41ca-9623-243b811b3491-kube-api-access-4zwb6\") pod \"barbican-operator-controller-manager-7d9dfd778-99gdz\" (UID: \"3a81ced4-fe81-41ca-9623-243b811b3491\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.734437 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tnbx\" (UniqueName: \"kubernetes.io/projected/2fd92e49-8ea8-44c1-b516-a47c2c9bc516-kube-api-access-6tnbx\") pod \"cinder-operator-controller-manager-6c677c69b-kkcfl\" (UID: \"2fd92e49-8ea8-44c1-b516-a47c2c9bc516\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.734474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9h2x\" (UniqueName: \"kubernetes.io/projected/f0669fb0-6164-4e2a-8f38-94d3a288923e-kube-api-access-p9h2x\") pod \"horizon-operator-controller-manager-68c6d99b8f-6fvml\" (UID: \"f0669fb0-6164-4e2a-8f38-94d3a288923e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.734491 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2l2j\" (UniqueName: \"kubernetes.io/projected/6a960bb0-9f2c-4610-a418-791038fd1a06-kube-api-access-d2l2j\") pod \"glance-operator-controller-manager-5697bb5779-xb72b\" (UID: \"6a960bb0-9f2c-4610-a418-791038fd1a06\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.734509 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhzcm\" (UniqueName: \"kubernetes.io/projected/badfd685-2701-4ced-a77e-6eb7b7ff8c01-kube-api-access-qhzcm\") pod \"designate-operator-controller-manager-697fb699cf-c7qtj\" (UID: \"badfd685-2701-4ced-a77e-6eb7b7ff8c01\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.737534 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-gzbs8" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.742197 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.746670 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.766498 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.770338 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-kdmnz" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.777546 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.792296 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.800792 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.802652 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zwb6\" (UniqueName: \"kubernetes.io/projected/3a81ced4-fe81-41ca-9623-243b811b3491-kube-api-access-4zwb6\") pod \"barbican-operator-controller-manager-7d9dfd778-99gdz\" (UID: \"3a81ced4-fe81-41ca-9623-243b811b3491\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.803581 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nbprk" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.803726 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.804834 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.806561 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2l2j\" (UniqueName: \"kubernetes.io/projected/6a960bb0-9f2c-4610-a418-791038fd1a06-kube-api-access-d2l2j\") pod \"glance-operator-controller-manager-5697bb5779-xb72b\" (UID: \"6a960bb0-9f2c-4610-a418-791038fd1a06\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.807224 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-gs7wj" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.809328 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fsqq\" (UniqueName: \"kubernetes.io/projected/1a65ee6c-f87e-4183-8c8c-9582347febda-kube-api-access-6fsqq\") pod \"heat-operator-controller-manager-5f64f6f8bb-lq68m\" (UID: \"1a65ee6c-f87e-4183-8c8c-9582347febda\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.819135 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhzcm\" (UniqueName: \"kubernetes.io/projected/badfd685-2701-4ced-a77e-6eb7b7ff8c01-kube-api-access-qhzcm\") pod \"designate-operator-controller-manager-697fb699cf-c7qtj\" (UID: \"badfd685-2701-4ced-a77e-6eb7b7ff8c01\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.821203 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tnbx\" (UniqueName: \"kubernetes.io/projected/2fd92e49-8ea8-44c1-b516-a47c2c9bc516-kube-api-access-6tnbx\") pod \"cinder-operator-controller-manager-6c677c69b-kkcfl\" (UID: \"2fd92e49-8ea8-44c1-b516-a47c2c9bc516\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.821277 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.827191 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.833066 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835084 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9wzc\" (UniqueName: \"kubernetes.io/projected/ea0450b1-401e-42af-9e33-f29b60af26ed-kube-api-access-q9wzc\") pod \"keystone-operator-controller-manager-7765d96ddf-wphkl\" (UID: \"ea0450b1-401e-42af-9e33-f29b60af26ed\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9h2x\" (UniqueName: \"kubernetes.io/projected/f0669fb0-6164-4e2a-8f38-94d3a288923e-kube-api-access-p9h2x\") pod \"horizon-operator-controller-manager-68c6d99b8f-6fvml\" (UID: \"f0669fb0-6164-4e2a-8f38-94d3a288923e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835207 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqkh5\" (UniqueName: \"kubernetes.io/projected/119c37b8-34ea-403c-9d87-6f252ff9a983-kube-api-access-cqkh5\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835236 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69b62\" (UniqueName: \"kubernetes.io/projected/df0a10a8-f92d-4cd9-843c-8769455c9b24-kube-api-access-69b62\") pod \"ironic-operator-controller-manager-967d97867-6fk8l\" (UID: \"df0a10a8-f92d-4cd9-843c-8769455c9b24\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835320 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsb5f\" (UniqueName: \"kubernetes.io/projected/1ca34247-7826-4547-9167-b6357ce63c62-kube-api-access-nsb5f\") pod \"manila-operator-controller-manager-5b5fd79c9c-jlzvg\" (UID: \"1ca34247-7826-4547-9167-b6357ce63c62\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.835345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k52fm\" (UniqueName: \"kubernetes.io/projected/608e5221-67cf-4e69-9ae1-037b120ff932-kube-api-access-k52fm\") pod \"mariadb-operator-controller-manager-79c8c4686c-kclgx\" (UID: \"608e5221-67cf-4e69-9ae1-037b120ff932\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.859228 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.878468 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.894621 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.909647 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9h2x\" (UniqueName: \"kubernetes.io/projected/f0669fb0-6164-4e2a-8f38-94d3a288923e-kube-api-access-p9h2x\") pod \"horizon-operator-controller-manager-68c6d99b8f-6fvml\" (UID: \"f0669fb0-6164-4e2a-8f38-94d3a288923e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.928528 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.935612 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg"] Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936269 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsb5f\" (UniqueName: \"kubernetes.io/projected/1ca34247-7826-4547-9167-b6357ce63c62-kube-api-access-nsb5f\") pod \"manila-operator-controller-manager-5b5fd79c9c-jlzvg\" (UID: \"1ca34247-7826-4547-9167-b6357ce63c62\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936314 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k52fm\" (UniqueName: \"kubernetes.io/projected/608e5221-67cf-4e69-9ae1-037b120ff932-kube-api-access-k52fm\") pod \"mariadb-operator-controller-manager-79c8c4686c-kclgx\" (UID: \"608e5221-67cf-4e69-9ae1-037b120ff932\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936362 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9wzc\" (UniqueName: \"kubernetes.io/projected/ea0450b1-401e-42af-9e33-f29b60af26ed-kube-api-access-q9wzc\") pod \"keystone-operator-controller-manager-7765d96ddf-wphkl\" (UID: \"ea0450b1-401e-42af-9e33-f29b60af26ed\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936440 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqkh5\" (UniqueName: \"kubernetes.io/projected/119c37b8-34ea-403c-9d87-6f252ff9a983-kube-api-access-cqkh5\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936514 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69b62\" (UniqueName: \"kubernetes.io/projected/df0a10a8-f92d-4cd9-843c-8769455c9b24-kube-api-access-69b62\") pod \"ironic-operator-controller-manager-967d97867-6fk8l\" (UID: \"df0a10a8-f92d-4cd9-843c-8769455c9b24\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.936678 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:12:52 crc kubenswrapper[4894]: E1210 19:12:52.937162 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:52 crc kubenswrapper[4894]: E1210 19:12:52.937215 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert podName:119c37b8-34ea-403c-9d87-6f252ff9a983 nodeName:}" failed. No retries permitted until 2025-12-10 19:12:53.43719623 +0000 UTC m=+1056.232044022 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert") pod "infra-operator-controller-manager-78d48bff9d-wqv7b" (UID: "119c37b8-34ea-403c-9d87-6f252ff9a983") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.958350 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.987569 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-68pql" Dec 10 19:12:52 crc kubenswrapper[4894]: I1210 19:12:52.997306 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:52.998716 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.008428 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k52fm\" (UniqueName: \"kubernetes.io/projected/608e5221-67cf-4e69-9ae1-037b120ff932-kube-api-access-k52fm\") pod \"mariadb-operator-controller-manager-79c8c4686c-kclgx\" (UID: \"608e5221-67cf-4e69-9ae1-037b120ff932\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.018740 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-qmfxt" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.018788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9wzc\" (UniqueName: \"kubernetes.io/projected/ea0450b1-401e-42af-9e33-f29b60af26ed-kube-api-access-q9wzc\") pod \"keystone-operator-controller-manager-7765d96ddf-wphkl\" (UID: \"ea0450b1-401e-42af-9e33-f29b60af26ed\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.018874 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.020202 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69b62\" (UniqueName: \"kubernetes.io/projected/df0a10a8-f92d-4cd9-843c-8769455c9b24-kube-api-access-69b62\") pod \"ironic-operator-controller-manager-967d97867-6fk8l\" (UID: \"df0a10a8-f92d-4cd9-843c-8769455c9b24\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.037057 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.037838 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s72zm\" (UniqueName: \"kubernetes.io/projected/cc54976b-bc17-4cb5-8eb3-901a163c2ec8-kube-api-access-s72zm\") pod \"nova-operator-controller-manager-697bc559fc-4dvmk\" (UID: \"cc54976b-bc17-4cb5-8eb3-901a163c2ec8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.037965 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbpv7\" (UniqueName: \"kubernetes.io/projected/0f98c7f8-f73d-4ca7-8361-91a719ab74ae-kube-api-access-mbpv7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2j6cg\" (UID: \"0f98c7f8-f73d-4ca7-8361-91a719ab74ae\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.045271 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsb5f\" (UniqueName: \"kubernetes.io/projected/1ca34247-7826-4547-9167-b6357ce63c62-kube-api-access-nsb5f\") pod \"manila-operator-controller-manager-5b5fd79c9c-jlzvg\" (UID: \"1ca34247-7826-4547-9167-b6357ce63c62\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.053800 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.059382 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqkh5\" (UniqueName: \"kubernetes.io/projected/119c37b8-34ea-403c-9d87-6f252ff9a983-kube-api-access-cqkh5\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.075222 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.076324 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.079296 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.081298 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8xvpq" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.082284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.140909 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.142306 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.143646 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbpv7\" (UniqueName: \"kubernetes.io/projected/0f98c7f8-f73d-4ca7-8361-91a719ab74ae-kube-api-access-mbpv7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2j6cg\" (UID: \"0f98c7f8-f73d-4ca7-8361-91a719ab74ae\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.143690 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkz4r\" (UniqueName: \"kubernetes.io/projected/e1ef63d6-b70d-4424-86a3-988318cb4241-kube-api-access-zkz4r\") pod \"octavia-operator-controller-manager-998648c74-qsjsk\" (UID: \"e1ef63d6-b70d-4424-86a3-988318cb4241\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.143782 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s72zm\" (UniqueName: \"kubernetes.io/projected/cc54976b-bc17-4cb5-8eb3-901a163c2ec8-kube-api-access-s72zm\") pod \"nova-operator-controller-manager-697bc559fc-4dvmk\" (UID: \"cc54976b-bc17-4cb5-8eb3-901a163c2ec8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.149023 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-r42xb" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.215161 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.216597 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.245244 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xtknb" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.245418 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.261480 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.264353 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkz4r\" (UniqueName: \"kubernetes.io/projected/e1ef63d6-b70d-4424-86a3-988318cb4241-kube-api-access-zkz4r\") pod \"octavia-operator-controller-manager-998648c74-qsjsk\" (UID: \"e1ef63d6-b70d-4424-86a3-988318cb4241\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.264394 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wk82\" (UniqueName: \"kubernetes.io/projected/1b2f13e5-e186-439b-8c83-fbdb422b567b-kube-api-access-9wk82\") pod \"ovn-operator-controller-manager-b6456fdb6-cg4zq\" (UID: \"1b2f13e5-e186-439b-8c83-fbdb422b567b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.264441 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.264468 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nzl4\" (UniqueName: \"kubernetes.io/projected/b3bce49e-40c1-4e29-9905-825760e0f21d-kube-api-access-7nzl4\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.278076 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.280559 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbpv7\" (UniqueName: \"kubernetes.io/projected/0f98c7f8-f73d-4ca7-8361-91a719ab74ae-kube-api-access-mbpv7\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2j6cg\" (UID: \"0f98c7f8-f73d-4ca7-8361-91a719ab74ae\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.280603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s72zm\" (UniqueName: \"kubernetes.io/projected/cc54976b-bc17-4cb5-8eb3-901a163c2ec8-kube-api-access-s72zm\") pod \"nova-operator-controller-manager-697bc559fc-4dvmk\" (UID: \"cc54976b-bc17-4cb5-8eb3-901a163c2ec8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.282065 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.304356 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.308510 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.308612 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.318180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.318368 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-k9qlp" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.366706 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkz4r\" (UniqueName: \"kubernetes.io/projected/e1ef63d6-b70d-4424-86a3-988318cb4241-kube-api-access-zkz4r\") pod \"octavia-operator-controller-manager-998648c74-qsjsk\" (UID: \"e1ef63d6-b70d-4424-86a3-988318cb4241\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.373543 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.381035 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wk82\" (UniqueName: \"kubernetes.io/projected/1b2f13e5-e186-439b-8c83-fbdb422b567b-kube-api-access-9wk82\") pod \"ovn-operator-controller-manager-b6456fdb6-cg4zq\" (UID: \"1b2f13e5-e186-439b-8c83-fbdb422b567b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.381120 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v6c5\" (UniqueName: \"kubernetes.io/projected/08615ae3-02b0-41fb-8951-433da5904be5-kube-api-access-2v6c5\") pod \"placement-operator-controller-manager-78f8948974-4dkpz\" (UID: \"08615ae3-02b0-41fb-8951-433da5904be5\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.381186 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.381239 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nzl4\" (UniqueName: \"kubernetes.io/projected/b3bce49e-40c1-4e29-9905-825760e0f21d-kube-api-access-7nzl4\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: E1210 19:12:53.382424 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:53 crc kubenswrapper[4894]: E1210 19:12:53.382497 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert podName:b3bce49e-40c1-4e29-9905-825760e0f21d nodeName:}" failed. No retries permitted until 2025-12-10 19:12:53.882480441 +0000 UTC m=+1056.677328233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert") pod "openstack-baremetal-operator-controller-manager-84b575879frxskr" (UID: "b3bce49e-40c1-4e29-9905-825760e0f21d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.482270 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v6c5\" (UniqueName: \"kubernetes.io/projected/08615ae3-02b0-41fb-8951-433da5904be5-kube-api-access-2v6c5\") pod \"placement-operator-controller-manager-78f8948974-4dkpz\" (UID: \"08615ae3-02b0-41fb-8951-433da5904be5\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.482395 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:53 crc kubenswrapper[4894]: E1210 19:12:53.482551 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:53 crc kubenswrapper[4894]: E1210 19:12:53.482604 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert podName:119c37b8-34ea-403c-9d87-6f252ff9a983 nodeName:}" failed. No retries permitted until 2025-12-10 19:12:54.482585723 +0000 UTC m=+1057.277433515 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert") pod "infra-operator-controller-manager-78d48bff9d-wqv7b" (UID: "119c37b8-34ea-403c-9d87-6f252ff9a983") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.518132 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wk82\" (UniqueName: \"kubernetes.io/projected/1b2f13e5-e186-439b-8c83-fbdb422b567b-kube-api-access-9wk82\") pod \"ovn-operator-controller-manager-b6456fdb6-cg4zq\" (UID: \"1b2f13e5-e186-439b-8c83-fbdb422b567b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.518475 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.537309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v6c5\" (UniqueName: \"kubernetes.io/projected/08615ae3-02b0-41fb-8951-433da5904be5-kube-api-access-2v6c5\") pod \"placement-operator-controller-manager-78f8948974-4dkpz\" (UID: \"08615ae3-02b0-41fb-8951-433da5904be5\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.538597 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nzl4\" (UniqueName: \"kubernetes.io/projected/b3bce49e-40c1-4e29-9905-825760e0f21d-kube-api-access-7nzl4\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.588245 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.591655 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.597042 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.600194 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-zktr8" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.651436 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.654071 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.678052 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.694838 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.696097 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.704678 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-j8r6f" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.731624 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.743866 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.743984 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.746865 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gwdpb" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.763523 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.790224 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk4r9\" (UniqueName: \"kubernetes.io/projected/ddfceacd-628c-4f2f-bdc5-2c8cd7329098-kube-api-access-kk4r9\") pod \"swift-operator-controller-manager-9d58d64bc-c96bw\" (UID: \"ddfceacd-628c-4f2f-bdc5-2c8cd7329098\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.790422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cszr\" (UniqueName: \"kubernetes.io/projected/7dd9f4f8-92dc-496a-a9b1-647a1619b7a5-kube-api-access-6cszr\") pod \"telemetry-operator-controller-manager-954d6f856-m8wb7\" (UID: \"7dd9f4f8-92dc-496a-a9b1-647a1619b7a5\") " pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.813256 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.815028 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.817351 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6qw6p" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.823878 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh"] Dec 10 19:12:53 crc kubenswrapper[4894]: W1210 19:12:53.848222 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fd92e49_8ea8_44c1_b516_a47c2c9bc516.slice/crio-1e52f7d9d28b7301e90d2af4ad26a155f4a4fd1b8a3b7b325ea15febd3576827 WatchSource:0}: Error finding container 1e52f7d9d28b7301e90d2af4ad26a155f4a4fd1b8a3b7b325ea15febd3576827: Status 404 returned error can't find the container with id 1e52f7d9d28b7301e90d2af4ad26a155f4a4fd1b8a3b7b325ea15febd3576827 Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.854360 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.855246 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.864682 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.864894 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.865095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-st9lw" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.890265 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.892731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.892809 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbk7h\" (UniqueName: \"kubernetes.io/projected/bcd78868-eb7e-4dfd-8992-2dd463f60287-kube-api-access-rbk7h\") pod \"watcher-operator-controller-manager-75944c9b7-qlq67\" (UID: \"bcd78868-eb7e-4dfd-8992-2dd463f60287\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.892948 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctxwj\" (UniqueName: \"kubernetes.io/projected/74c8428d-0015-4712-8bfa-6bf12f68b0bf-kube-api-access-ctxwj\") pod \"test-operator-controller-manager-5854674fcc-rmkmh\" (UID: \"74c8428d-0015-4712-8bfa-6bf12f68b0bf\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.892998 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk4r9\" (UniqueName: \"kubernetes.io/projected/ddfceacd-628c-4f2f-bdc5-2c8cd7329098-kube-api-access-kk4r9\") pod \"swift-operator-controller-manager-9d58d64bc-c96bw\" (UID: \"ddfceacd-628c-4f2f-bdc5-2c8cd7329098\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.893149 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cszr\" (UniqueName: \"kubernetes.io/projected/7dd9f4f8-92dc-496a-a9b1-647a1619b7a5-kube-api-access-6cszr\") pod \"telemetry-operator-controller-manager-954d6f856-m8wb7\" (UID: \"7dd9f4f8-92dc-496a-a9b1-647a1619b7a5\") " pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:12:53 crc kubenswrapper[4894]: E1210 19:12:53.893559 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:53 crc kubenswrapper[4894]: E1210 19:12:53.893606 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert podName:b3bce49e-40c1-4e29-9905-825760e0f21d nodeName:}" failed. No retries permitted until 2025-12-10 19:12:54.893590703 +0000 UTC m=+1057.688438495 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert") pod "openstack-baremetal-operator-controller-manager-84b575879frxskr" (UID: "b3bce49e-40c1-4e29-9905-825760e0f21d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.910574 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.918607 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.934497 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.936178 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl"] Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.941066 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk4r9\" (UniqueName: \"kubernetes.io/projected/ddfceacd-628c-4f2f-bdc5-2c8cd7329098-kube-api-access-kk4r9\") pod \"swift-operator-controller-manager-9d58d64bc-c96bw\" (UID: \"ddfceacd-628c-4f2f-bdc5-2c8cd7329098\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.941155 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-7mc8d" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.955455 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cszr\" (UniqueName: \"kubernetes.io/projected/7dd9f4f8-92dc-496a-a9b1-647a1619b7a5-kube-api-access-6cszr\") pod \"telemetry-operator-controller-manager-954d6f856-m8wb7\" (UID: \"7dd9f4f8-92dc-496a-a9b1-647a1619b7a5\") " pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.969059 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.994834 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctxwj\" (UniqueName: \"kubernetes.io/projected/74c8428d-0015-4712-8bfa-6bf12f68b0bf-kube-api-access-ctxwj\") pod \"test-operator-controller-manager-5854674fcc-rmkmh\" (UID: \"74c8428d-0015-4712-8bfa-6bf12f68b0bf\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.994913 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.994989 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx699\" (UniqueName: \"kubernetes.io/projected/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-kube-api-access-qx699\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.995038 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.995114 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5842\" (UniqueName: \"kubernetes.io/projected/e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af-kube-api-access-d5842\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mzm6b\" (UID: \"e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" Dec 10 19:12:53 crc kubenswrapper[4894]: I1210 19:12:53.995185 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbk7h\" (UniqueName: \"kubernetes.io/projected/bcd78868-eb7e-4dfd-8992-2dd463f60287-kube-api-access-rbk7h\") pod \"watcher-operator-controller-manager-75944c9b7-qlq67\" (UID: \"bcd78868-eb7e-4dfd-8992-2dd463f60287\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.014515 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.021318 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbk7h\" (UniqueName: \"kubernetes.io/projected/bcd78868-eb7e-4dfd-8992-2dd463f60287-kube-api-access-rbk7h\") pod \"watcher-operator-controller-manager-75944c9b7-qlq67\" (UID: \"bcd78868-eb7e-4dfd-8992-2dd463f60287\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.026155 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctxwj\" (UniqueName: \"kubernetes.io/projected/74c8428d-0015-4712-8bfa-6bf12f68b0bf-kube-api-access-ctxwj\") pod \"test-operator-controller-manager-5854674fcc-rmkmh\" (UID: \"74c8428d-0015-4712-8bfa-6bf12f68b0bf\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.067977 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.083675 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.104368 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5842\" (UniqueName: \"kubernetes.io/projected/e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af-kube-api-access-d5842\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mzm6b\" (UID: \"e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.104473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.104512 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx699\" (UniqueName: \"kubernetes.io/projected/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-kube-api-access-qx699\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.104547 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.104676 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.104730 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:12:54.604710725 +0000 UTC m=+1057.399558517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "metrics-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.104894 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.104946 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:12:54.604928892 +0000 UTC m=+1057.399776684 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.143951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx699\" (UniqueName: \"kubernetes.io/projected/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-kube-api-access-qx699\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.147588 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5842\" (UniqueName: \"kubernetes.io/projected/e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af-kube-api-access-d5842\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mzm6b\" (UID: \"e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.211346 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.217158 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.288725 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.305138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.433284 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" event={"ID":"6a960bb0-9f2c-4610-a418-791038fd1a06","Type":"ContainerStarted","Data":"9ead52ed6e99f79bf0f09caedb8155ace2cff6a827cf92ec4ee3bb4bc603f58d"} Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.449392 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" event={"ID":"badfd685-2701-4ced-a77e-6eb7b7ff8c01","Type":"ContainerStarted","Data":"5a6dd31acabd30ba160a1de9838efe3d599ef484bd77d3c27902c84c8e9cdca7"} Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.450060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" event={"ID":"3a81ced4-fe81-41ca-9623-243b811b3491","Type":"ContainerStarted","Data":"3d53a82dd4a28ffa77d3576d3a2afc456633a53722716d9d704232f72a304278"} Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.450635 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" event={"ID":"2fd92e49-8ea8-44c1-b516-a47c2c9bc516","Type":"ContainerStarted","Data":"1e52f7d9d28b7301e90d2af4ad26a155f4a4fd1b8a3b7b325ea15febd3576827"} Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.488290 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.523077 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.523303 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.523348 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert podName:119c37b8-34ea-403c-9d87-6f252ff9a983 nodeName:}" failed. No retries permitted until 2025-12-10 19:12:56.523332581 +0000 UTC m=+1059.318180373 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert") pod "infra-operator-controller-manager-78d48bff9d-wqv7b" (UID: "119c37b8-34ea-403c-9d87-6f252ff9a983") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: W1210 19:12:54.529126 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a65ee6c_f87e_4183_8c8c_9582347febda.slice/crio-f3576451ffbd3eda6d6476ce55d1443127dffd89b2bc18b420df5a0ee5302e0f WatchSource:0}: Error finding container f3576451ffbd3eda6d6476ce55d1443127dffd89b2bc18b420df5a0ee5302e0f: Status 404 returned error can't find the container with id f3576451ffbd3eda6d6476ce55d1443127dffd89b2bc18b420df5a0ee5302e0f Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.624635 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.625131 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.625233 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.625322 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.625330 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:12:55.625305235 +0000 UTC m=+1058.420153027 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "metrics-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.625369 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:12:55.625359356 +0000 UTC m=+1058.420207248 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.734088 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.748906 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l"] Dec 10 19:12:54 crc kubenswrapper[4894]: W1210 19:12:54.749049 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod608e5221_67cf_4e69_9ae1_037b120ff932.slice/crio-dc060e606811018c31ba819b2548b52e5c368e84bf3c948504e2c828ac5f8f14 WatchSource:0}: Error finding container dc060e606811018c31ba819b2548b52e5c368e84bf3c948504e2c828ac5f8f14: Status 404 returned error can't find the container with id dc060e606811018c31ba819b2548b52e5c368e84bf3c948504e2c828ac5f8f14 Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.840592 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.857669 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.868444 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.928700 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.928904 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: E1210 19:12:54.929021 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert podName:b3bce49e-40c1-4e29-9905-825760e0f21d nodeName:}" failed. No retries permitted until 2025-12-10 19:12:56.929001495 +0000 UTC m=+1059.723849287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert") pod "openstack-baremetal-operator-controller-manager-84b575879frxskr" (UID: "b3bce49e-40c1-4e29-9905-825760e0f21d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.951004 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.959520 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg"] Dec 10 19:12:54 crc kubenswrapper[4894]: W1210 19:12:54.964629 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1ef63d6_b70d_4424_86a3_988318cb4241.slice/crio-3a417e0153bda35aed77b6993639cff224492374d5b6da1e1829ddd4b45c3e3b WatchSource:0}: Error finding container 3a417e0153bda35aed77b6993639cff224492374d5b6da1e1829ddd4b45c3e3b: Status 404 returned error can't find the container with id 3a417e0153bda35aed77b6993639cff224492374d5b6da1e1829ddd4b45c3e3b Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.967691 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.975685 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz"] Dec 10 19:12:54 crc kubenswrapper[4894]: I1210 19:12:54.981480 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg"] Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.162242 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw"] Dec 10 19:12:55 crc kubenswrapper[4894]: W1210 19:12:55.168097 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddfceacd_628c_4f2f_bdc5_2c8cd7329098.slice/crio-9ceb4a27301d9c1ffabaf2b60d035a0fed18cb97b7f5eacbfb018d0236408f80 WatchSource:0}: Error finding container 9ceb4a27301d9c1ffabaf2b60d035a0fed18cb97b7f5eacbfb018d0236408f80: Status 404 returned error can't find the container with id 9ceb4a27301d9c1ffabaf2b60d035a0fed18cb97b7f5eacbfb018d0236408f80 Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.168901 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq"] Dec 10 19:12:55 crc kubenswrapper[4894]: W1210 19:12:55.169309 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1fb825a_ebc5_4cdf_949c_83d9bfe6c5af.slice/crio-a7ab6dee00f0f0e25b621f29c35723b3b9a5022ee8f63a017ecf75abbe8a1e44 WatchSource:0}: Error finding container a7ab6dee00f0f0e25b621f29c35723b3b9a5022ee8f63a017ecf75abbe8a1e44: Status 404 returned error can't find the container with id a7ab6dee00f0f0e25b621f29c35723b3b9a5022ee8f63a017ecf75abbe8a1e44 Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.173052 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d5842,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-mzm6b_openstack-operators(e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.174277 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" podUID="e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.178143 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b"] Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.184598 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67"] Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.193220 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9wk82,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-cg4zq_openstack-operators(1b2f13e5-e186-439b-8c83-fbdb422b567b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.196633 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9wk82,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-cg4zq_openstack-operators(1b2f13e5-e186-439b-8c83-fbdb422b567b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.197718 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" podUID="1b2f13e5-e186-439b-8c83-fbdb422b567b" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.198154 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh"] Dec 10 19:12:55 crc kubenswrapper[4894]: W1210 19:12:55.198229 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcd78868_eb7e_4dfd_8992_2dd463f60287.slice/crio-aa3d076e7292c967a24119f7f89ac2b4208d28c2800a3e43025836d4f54cbf8a WatchSource:0}: Error finding container aa3d076e7292c967a24119f7f89ac2b4208d28c2800a3e43025836d4f54cbf8a: Status 404 returned error can't find the container with id aa3d076e7292c967a24119f7f89ac2b4208d28c2800a3e43025836d4f54cbf8a Dec 10 19:12:55 crc kubenswrapper[4894]: W1210 19:12:55.199244 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74c8428d_0015_4712_8bfa_6bf12f68b0bf.slice/crio-ba024f15108d740854486802e89bc84fac7e61757cf9016e2c4870d9aa90576b WatchSource:0}: Error finding container ba024f15108d740854486802e89bc84fac7e61757cf9016e2c4870d9aa90576b: Status 404 returned error can't find the container with id ba024f15108d740854486802e89bc84fac7e61757cf9016e2c4870d9aa90576b Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.201497 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ctxwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-rmkmh_openstack-operators(74c8428d-0015-4712-8bfa-6bf12f68b0bf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.202309 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rbk7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-qlq67_openstack-operators(bcd78868-eb7e-4dfd-8992-2dd463f60287): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.205008 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rbk7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-qlq67_openstack-operators(bcd78868-eb7e-4dfd-8992-2dd463f60287): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.205186 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ctxwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-rmkmh_openstack-operators(74c8428d-0015-4712-8bfa-6bf12f68b0bf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.206523 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" podUID="74c8428d-0015-4712-8bfa-6bf12f68b0bf" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.206582 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" podUID="bcd78868-eb7e-4dfd-8992-2dd463f60287" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.462616 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" event={"ID":"08615ae3-02b0-41fb-8951-433da5904be5","Type":"ContainerStarted","Data":"8e06705c333d4d11bd5c5d5f33274af4a0314fe2bd70c2c4e4286414247df348"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.463576 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" event={"ID":"df0a10a8-f92d-4cd9-843c-8769455c9b24","Type":"ContainerStarted","Data":"5633a05291ed6e4eafbf985aec61fe18cf836e6d1478e287030ae06959f7f703"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.464619 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" event={"ID":"bcd78868-eb7e-4dfd-8992-2dd463f60287","Type":"ContainerStarted","Data":"aa3d076e7292c967a24119f7f89ac2b4208d28c2800a3e43025836d4f54cbf8a"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.466929 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" event={"ID":"74c8428d-0015-4712-8bfa-6bf12f68b0bf","Type":"ContainerStarted","Data":"ba024f15108d740854486802e89bc84fac7e61757cf9016e2c4870d9aa90576b"} Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.467460 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" podUID="bcd78868-eb7e-4dfd-8992-2dd463f60287" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.474758 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" podUID="74c8428d-0015-4712-8bfa-6bf12f68b0bf" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.477021 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" event={"ID":"cc54976b-bc17-4cb5-8eb3-901a163c2ec8","Type":"ContainerStarted","Data":"e7821a21ea45cb3bb2dac1cbaa502beb23790973a8ba184d6774b420e6918786"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.478249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" event={"ID":"f0669fb0-6164-4e2a-8f38-94d3a288923e","Type":"ContainerStarted","Data":"931112de780e068b1bb71e853a2e66f8c23f97c4c858281f920d73aa0c6ef025"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.484942 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" event={"ID":"7dd9f4f8-92dc-496a-a9b1-647a1619b7a5","Type":"ContainerStarted","Data":"922457d7bec6ce4e192b3f450153fafe97a106c7f89293bbd90d0994416eb548"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.491032 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" event={"ID":"ddfceacd-628c-4f2f-bdc5-2c8cd7329098","Type":"ContainerStarted","Data":"9ceb4a27301d9c1ffabaf2b60d035a0fed18cb97b7f5eacbfb018d0236408f80"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.503031 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" event={"ID":"0f98c7f8-f73d-4ca7-8361-91a719ab74ae","Type":"ContainerStarted","Data":"eda70e80030e0a2a2150f69ee762ae37bf1105492abecb6eb7b9f353ef1beb39"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.508983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" event={"ID":"e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af","Type":"ContainerStarted","Data":"a7ab6dee00f0f0e25b621f29c35723b3b9a5022ee8f63a017ecf75abbe8a1e44"} Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.511252 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" podUID="e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.512962 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" event={"ID":"608e5221-67cf-4e69-9ae1-037b120ff932","Type":"ContainerStarted","Data":"dc060e606811018c31ba819b2548b52e5c368e84bf3c948504e2c828ac5f8f14"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.514430 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" event={"ID":"1b2f13e5-e186-439b-8c83-fbdb422b567b","Type":"ContainerStarted","Data":"e2488cdd1d3c6d9621b0d412da36b5367f9fe41dbdc45258b3ad418600313775"} Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.517150 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" podUID="1b2f13e5-e186-439b-8c83-fbdb422b567b" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.517724 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" event={"ID":"ea0450b1-401e-42af-9e33-f29b60af26ed","Type":"ContainerStarted","Data":"b2d97bd836f7d625d5bdaa02a38e80b93329ec6a4492c0bbc5b58274d6ee6594"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.530681 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" event={"ID":"1a65ee6c-f87e-4183-8c8c-9582347febda","Type":"ContainerStarted","Data":"f3576451ffbd3eda6d6476ce55d1443127dffd89b2bc18b420df5a0ee5302e0f"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.535169 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" event={"ID":"1ca34247-7826-4547-9167-b6357ce63c62","Type":"ContainerStarted","Data":"91981fd9ecd8e269e9549fee6979136b808e16a31efb4df9b6aba7b9132d1a42"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.553710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" event={"ID":"e1ef63d6-b70d-4424-86a3-988318cb4241","Type":"ContainerStarted","Data":"3a417e0153bda35aed77b6993639cff224492374d5b6da1e1829ddd4b45c3e3b"} Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.645284 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:55 crc kubenswrapper[4894]: I1210 19:12:55.645375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.645469 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.645544 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:12:57.645514724 +0000 UTC m=+1060.440362516 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "webhook-server-cert" not found Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.646478 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:12:55 crc kubenswrapper[4894]: E1210 19:12:55.646547 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:12:57.646533232 +0000 UTC m=+1060.441381014 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "metrics-server-cert" not found Dec 10 19:12:56 crc kubenswrapper[4894]: I1210 19:12:56.577007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.577450 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.577515 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert podName:119c37b8-34ea-403c-9d87-6f252ff9a983 nodeName:}" failed. No retries permitted until 2025-12-10 19:13:00.577496854 +0000 UTC m=+1063.372344646 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert") pod "infra-operator-controller-manager-78d48bff9d-wqv7b" (UID: "119c37b8-34ea-403c-9d87-6f252ff9a983") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.584570 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" podUID="e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af" Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.591090 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" podUID="bcd78868-eb7e-4dfd-8992-2dd463f60287" Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.591117 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" podUID="74c8428d-0015-4712-8bfa-6bf12f68b0bf" Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.591181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" podUID="1b2f13e5-e186-439b-8c83-fbdb422b567b" Dec 10 19:12:56 crc kubenswrapper[4894]: I1210 19:12:56.987876 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.988017 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:56 crc kubenswrapper[4894]: E1210 19:12:56.988060 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert podName:b3bce49e-40c1-4e29-9905-825760e0f21d nodeName:}" failed. No retries permitted until 2025-12-10 19:13:00.98804594 +0000 UTC m=+1063.782893732 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert") pod "openstack-baremetal-operator-controller-manager-84b575879frxskr" (UID: "b3bce49e-40c1-4e29-9905-825760e0f21d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:12:57 crc kubenswrapper[4894]: I1210 19:12:57.707887 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:57 crc kubenswrapper[4894]: I1210 19:12:57.708291 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:12:57 crc kubenswrapper[4894]: E1210 19:12:57.708783 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:12:57 crc kubenswrapper[4894]: E1210 19:12:57.708871 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:13:01.708854986 +0000 UTC m=+1064.503702778 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "webhook-server-cert" not found Dec 10 19:12:57 crc kubenswrapper[4894]: E1210 19:12:57.709030 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:12:57 crc kubenswrapper[4894]: E1210 19:12:57.709079 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:13:01.709063231 +0000 UTC m=+1064.503911023 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "metrics-server-cert" not found Dec 10 19:13:00 crc kubenswrapper[4894]: I1210 19:13:00.670773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:00 crc kubenswrapper[4894]: E1210 19:13:00.676031 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:13:00 crc kubenswrapper[4894]: E1210 19:13:00.676127 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert podName:119c37b8-34ea-403c-9d87-6f252ff9a983 nodeName:}" failed. No retries permitted until 2025-12-10 19:13:08.676088782 +0000 UTC m=+1071.470936564 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert") pod "infra-operator-controller-manager-78d48bff9d-wqv7b" (UID: "119c37b8-34ea-403c-9d87-6f252ff9a983") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:13:01 crc kubenswrapper[4894]: I1210 19:13:01.078208 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:13:01 crc kubenswrapper[4894]: E1210 19:13:01.078392 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:13:01 crc kubenswrapper[4894]: E1210 19:13:01.078459 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert podName:b3bce49e-40c1-4e29-9905-825760e0f21d nodeName:}" failed. No retries permitted until 2025-12-10 19:13:09.078440315 +0000 UTC m=+1071.873288107 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert") pod "openstack-baremetal-operator-controller-manager-84b575879frxskr" (UID: "b3bce49e-40c1-4e29-9905-825760e0f21d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:13:01 crc kubenswrapper[4894]: I1210 19:13:01.788291 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:01 crc kubenswrapper[4894]: E1210 19:13:01.788483 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:13:01 crc kubenswrapper[4894]: E1210 19:13:01.788806 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:13:09.788785477 +0000 UTC m=+1072.583633269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "metrics-server-cert" not found Dec 10 19:13:01 crc kubenswrapper[4894]: I1210 19:13:01.788833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:01 crc kubenswrapper[4894]: E1210 19:13:01.789005 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:13:01 crc kubenswrapper[4894]: E1210 19:13:01.789030 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:13:09.789023193 +0000 UTC m=+1072.583870985 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "webhook-server-cert" not found Dec 10 19:13:08 crc kubenswrapper[4894]: E1210 19:13:08.437189 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 10 19:13:08 crc kubenswrapper[4894]: E1210 19:13:08.438019 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2v6c5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-4dkpz_openstack-operators(08615ae3-02b0-41fb-8951-433da5904be5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:08 crc kubenswrapper[4894]: I1210 19:13:08.730698 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:08 crc kubenswrapper[4894]: E1210 19:13:08.732510 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:13:08 crc kubenswrapper[4894]: E1210 19:13:08.732570 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert podName:119c37b8-34ea-403c-9d87-6f252ff9a983 nodeName:}" failed. No retries permitted until 2025-12-10 19:13:24.732551482 +0000 UTC m=+1087.527399284 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert") pod "infra-operator-controller-manager-78d48bff9d-wqv7b" (UID: "119c37b8-34ea-403c-9d87-6f252ff9a983") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:13:09 crc kubenswrapper[4894]: I1210 19:13:09.136952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.137173 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.137247 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert podName:b3bce49e-40c1-4e29-9905-825760e0f21d nodeName:}" failed. No retries permitted until 2025-12-10 19:13:25.137226289 +0000 UTC m=+1087.932074091 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert") pod "openstack-baremetal-operator-controller-manager-84b575879frxskr" (UID: "b3bce49e-40c1-4e29-9905-825760e0f21d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.319776 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.319965 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhzcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-c7qtj_openstack-operators(badfd685-2701-4ced-a77e-6eb7b7ff8c01): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:09 crc kubenswrapper[4894]: I1210 19:13:09.848318 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.848494 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.848561 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:13:25.848544917 +0000 UTC m=+1088.643392709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "metrics-server-cert" not found Dec 10 19:13:09 crc kubenswrapper[4894]: I1210 19:13:09.848645 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.848720 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:13:09 crc kubenswrapper[4894]: E1210 19:13:09.848756 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs podName:5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd nodeName:}" failed. No retries permitted until 2025-12-10 19:13:25.848746292 +0000 UTC m=+1088.643594084 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-x9dnx" (UID: "5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd") : secret "webhook-server-cert" not found Dec 10 19:13:10 crc kubenswrapper[4894]: E1210 19:13:10.388249 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 10 19:13:10 crc kubenswrapper[4894]: E1210 19:13:10.388479 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4zwb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-99gdz_openstack-operators(3a81ced4-fe81-41ca-9623-243b811b3491): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:11 crc kubenswrapper[4894]: E1210 19:13:11.540327 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 10 19:13:11 crc kubenswrapper[4894]: E1210 19:13:11.540494 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zkz4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-qsjsk_openstack-operators(e1ef63d6-b70d-4424-86a3-988318cb4241): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:12 crc kubenswrapper[4894]: E1210 19:13:12.466455 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 10 19:13:12 crc kubenswrapper[4894]: E1210 19:13:12.466943 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6fsqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-lq68m_openstack-operators(1a65ee6c-f87e-4183-8c8c-9582347febda): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.012282 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.012479 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s72zm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-4dvmk_openstack-operators(cc54976b-bc17-4cb5-8eb3-901a163c2ec8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.580650 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.581156 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-69b62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-6fk8l_openstack-operators(df0a10a8-f92d-4cd9-843c-8769455c9b24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.673274 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.65:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.673341 4894 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.65:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1" Dec 10 19:13:13 crc kubenswrapper[4894]: E1210 19:13:13.674541 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.65:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cszr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-954d6f856-m8wb7_openstack-operators(7dd9f4f8-92dc-496a-a9b1-647a1619b7a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:14 crc kubenswrapper[4894]: E1210 19:13:14.459505 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 10 19:13:14 crc kubenswrapper[4894]: E1210 19:13:14.459988 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kk4r9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-c96bw_openstack-operators(ddfceacd-628c-4f2f-bdc5-2c8cd7329098): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:15 crc kubenswrapper[4894]: E1210 19:13:15.186969 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 10 19:13:15 crc kubenswrapper[4894]: E1210 19:13:15.187144 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mbpv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-2j6cg_openstack-operators(0f98c7f8-f73d-4ca7-8361-91a719ab74ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:16 crc kubenswrapper[4894]: E1210 19:13:16.498100 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 10 19:13:16 crc kubenswrapper[4894]: E1210 19:13:16.498306 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q9wzc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-wphkl_openstack-operators(ea0450b1-401e-42af-9e33-f29b60af26ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:13:20 crc kubenswrapper[4894]: I1210 19:13:20.855336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" event={"ID":"608e5221-67cf-4e69-9ae1-037b120ff932","Type":"ContainerStarted","Data":"ad4d757c61db0759ce7fc928d59b267bb9b488b29d8e1b8632888923ec2189db"} Dec 10 19:13:20 crc kubenswrapper[4894]: I1210 19:13:20.857302 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" event={"ID":"f0669fb0-6164-4e2a-8f38-94d3a288923e","Type":"ContainerStarted","Data":"c8924c20292a9043d200f20eccb12ca25c7a7c7632bb6081060cde55230bef17"} Dec 10 19:13:20 crc kubenswrapper[4894]: I1210 19:13:20.858875 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" event={"ID":"6a960bb0-9f2c-4610-a418-791038fd1a06","Type":"ContainerStarted","Data":"d589821b0c85d1b9c323fbddc8cb169a0bbd12f9f4237c851e9c5bedfdee7ed1"} Dec 10 19:13:20 crc kubenswrapper[4894]: I1210 19:13:20.861684 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" event={"ID":"2fd92e49-8ea8-44c1-b516-a47c2c9bc516","Type":"ContainerStarted","Data":"4f2360e149075becd7dcf8ce44215a2878d5188c193b22a9f32e083fda342b27"} Dec 10 19:13:20 crc kubenswrapper[4894]: I1210 19:13:20.862777 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" event={"ID":"1ca34247-7826-4547-9167-b6357ce63c62","Type":"ContainerStarted","Data":"296f7ad081e81e1cc748bc53035e874283cc5aa0d62f3e85349ba528f43d4b76"} Dec 10 19:13:21 crc kubenswrapper[4894]: I1210 19:13:21.870283 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" event={"ID":"74c8428d-0015-4712-8bfa-6bf12f68b0bf","Type":"ContainerStarted","Data":"fb0368fdbb5d8e8e3c5b965f12d8320294d45d04edf4c7f659273620a9ec2a66"} Dec 10 19:13:22 crc kubenswrapper[4894]: I1210 19:13:22.880419 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" event={"ID":"1b2f13e5-e186-439b-8c83-fbdb422b567b","Type":"ContainerStarted","Data":"ab184f8c8dab2026af414559ce6975bbae109f9d63be6ded0bfa76c5200813d4"} Dec 10 19:13:23 crc kubenswrapper[4894]: E1210 19:13:23.538350 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" podUID="3a81ced4-fe81-41ca-9623-243b811b3491" Dec 10 19:13:23 crc kubenswrapper[4894]: E1210 19:13:23.545285 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" podUID="cc54976b-bc17-4cb5-8eb3-901a163c2ec8" Dec 10 19:13:23 crc kubenswrapper[4894]: E1210 19:13:23.873141 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" podUID="badfd685-2701-4ced-a77e-6eb7b7ff8c01" Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.935932 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" event={"ID":"bcd78868-eb7e-4dfd-8992-2dd463f60287","Type":"ContainerStarted","Data":"115e327f188af575f53735e1c1c1ad13e2628226d3e47f14e200f4ed27e69703"} Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.935971 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" event={"ID":"bcd78868-eb7e-4dfd-8992-2dd463f60287","Type":"ContainerStarted","Data":"130d23b10ec767da08f6d4f73f5df9f64313f4188a35b12bdd0bb1fc86bae17a"} Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.936924 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.955437 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" event={"ID":"cc54976b-bc17-4cb5-8eb3-901a163c2ec8","Type":"ContainerStarted","Data":"8c9293618787c65947c3204706101bfb2a6854f2aa0cb763686eead12ea62bb4"} Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.961389 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" event={"ID":"2fd92e49-8ea8-44c1-b516-a47c2c9bc516","Type":"ContainerStarted","Data":"686ab93aebbb1279e876243a0d356f3c4fe6428bbce82aca0f9e01f2445ba398"} Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.962575 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.968787 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" podStartSLOduration=6.2627479 podStartE2EDuration="30.968763098s" podCreationTimestamp="2025-12-10 19:12:53 +0000 UTC" firstStartedPulling="2025-12-10 19:12:55.202215306 +0000 UTC m=+1057.997063098" lastFinishedPulling="2025-12-10 19:13:19.908230464 +0000 UTC m=+1082.703078296" observedRunningTime="2025-12-10 19:13:23.96038196 +0000 UTC m=+1086.755229752" watchObservedRunningTime="2025-12-10 19:13:23.968763098 +0000 UTC m=+1086.763610890" Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.977038 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" event={"ID":"1ca34247-7826-4547-9167-b6357ce63c62","Type":"ContainerStarted","Data":"ae3da38b683afe186abec0c9309fd07e4dd32f03d914b0ac164a089db6e55f19"} Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.977495 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:13:23 crc kubenswrapper[4894]: I1210 19:13:23.981637 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" event={"ID":"badfd685-2701-4ced-a77e-6eb7b7ff8c01","Type":"ContainerStarted","Data":"515d3a4afba1580eb645d8d46acf2cd7535f759ad71b78f80feef37da0b1fb73"} Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.000528 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" event={"ID":"3a81ced4-fe81-41ca-9623-243b811b3491","Type":"ContainerStarted","Data":"3278fc5b5fcc558e4affb852e69df1e22f0d71804a1ea325778cff024347e2a3"} Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.013765 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" podStartSLOduration=3.7054190030000003 podStartE2EDuration="32.013751351s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.987576528 +0000 UTC m=+1057.782424320" lastFinishedPulling="2025-12-10 19:13:23.295908876 +0000 UTC m=+1086.090756668" observedRunningTime="2025-12-10 19:13:24.010330818 +0000 UTC m=+1086.805178610" watchObservedRunningTime="2025-12-10 19:13:24.013751351 +0000 UTC m=+1086.808599143" Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.024339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" event={"ID":"e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af","Type":"ContainerStarted","Data":"73a62fcb37873cf5968eeeae3473ee53a9492feed743b179d5a337c6b81adeb6"} Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.035779 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" podStartSLOduration=2.44241582 podStartE2EDuration="32.03576194s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:53.871358068 +0000 UTC m=+1056.666205860" lastFinishedPulling="2025-12-10 19:13:23.464704188 +0000 UTC m=+1086.259551980" observedRunningTime="2025-12-10 19:13:24.032502551 +0000 UTC m=+1086.827350363" watchObservedRunningTime="2025-12-10 19:13:24.03576194 +0000 UTC m=+1086.830609732" Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.076753 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mzm6b" podStartSLOduration=6.221635852 podStartE2EDuration="31.076736075s" podCreationTimestamp="2025-12-10 19:12:53 +0000 UTC" firstStartedPulling="2025-12-10 19:12:55.17295844 +0000 UTC m=+1057.967806232" lastFinishedPulling="2025-12-10 19:13:20.028058663 +0000 UTC m=+1082.822906455" observedRunningTime="2025-12-10 19:13:24.075758158 +0000 UTC m=+1086.870605970" watchObservedRunningTime="2025-12-10 19:13:24.076736075 +0000 UTC m=+1086.871583877" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.122786 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" podUID="1a65ee6c-f87e-4183-8c8c-9582347febda" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.182297 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" podUID="df0a10a8-f92d-4cd9-843c-8769455c9b24" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.280289 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" podUID="ea0450b1-401e-42af-9e33-f29b60af26ed" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.332912 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" podUID="e1ef63d6-b70d-4424-86a3-988318cb4241" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.358188 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" podUID="0f98c7f8-f73d-4ca7-8361-91a719ab74ae" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.361823 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" podUID="7dd9f4f8-92dc-496a-a9b1-647a1619b7a5" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.383789 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" podUID="08615ae3-02b0-41fb-8951-433da5904be5" Dec 10 19:13:24 crc kubenswrapper[4894]: E1210 19:13:24.398311 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" podUID="ddfceacd-628c-4f2f-bdc5-2c8cd7329098" Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.804865 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.823631 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/119c37b8-34ea-403c-9d87-6f252ff9a983-cert\") pod \"infra-operator-controller-manager-78d48bff9d-wqv7b\" (UID: \"119c37b8-34ea-403c-9d87-6f252ff9a983\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.838510 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vx7jp" Dec 10 19:13:24 crc kubenswrapper[4894]: I1210 19:13:24.842766 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.049491 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" event={"ID":"608e5221-67cf-4e69-9ae1-037b120ff932","Type":"ContainerStarted","Data":"bc39eba5488c3928b2f35d75796cfab52123e67361438f2785b75e5226769f5d"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.050041 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.051556 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.058279 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" event={"ID":"1b2f13e5-e186-439b-8c83-fbdb422b567b","Type":"ContainerStarted","Data":"70b7a275a8ec96be000a209991ade64b8065225f799e318ebd4a459c9568b8e1"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.058418 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.059946 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" event={"ID":"ddfceacd-628c-4f2f-bdc5-2c8cd7329098","Type":"ContainerStarted","Data":"29c7bc70796fff83ba56d25a0f2ea6ff57d434fee770415d859f0742a32f2923"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.065576 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" event={"ID":"0f98c7f8-f73d-4ca7-8361-91a719ab74ae","Type":"ContainerStarted","Data":"e2ffa245c4c32d8096f6b5e4afeb61c131846eac2eb5bf8a9b58ccdc680d8847"} Dec 10 19:13:25 crc kubenswrapper[4894]: E1210 19:13:25.066795 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" podUID="0f98c7f8-f73d-4ca7-8361-91a719ab74ae" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.068482 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" event={"ID":"1a65ee6c-f87e-4183-8c8c-9582347febda","Type":"ContainerStarted","Data":"70e1dadf0a8be6fb10ff5ebd69476fbf58fcd02d5b74780871b7076ddd7a4984"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.070990 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" event={"ID":"f0669fb0-6164-4e2a-8f38-94d3a288923e","Type":"ContainerStarted","Data":"c7911bf6d4901aaa7064fccc44f22a22d67ac6f47a717ccfab0d5d49355ef095"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.072239 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.076658 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.083596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" event={"ID":"e1ef63d6-b70d-4424-86a3-988318cb4241","Type":"ContainerStarted","Data":"e88cd9cf20bc1be970495b69cd943d9a1ef910ce487f5ee69554373150bdaea1"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.087154 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-kclgx" podStartSLOduration=4.140327162 podStartE2EDuration="33.087126416s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.751284311 +0000 UTC m=+1057.546132103" lastFinishedPulling="2025-12-10 19:13:23.698083565 +0000 UTC m=+1086.492931357" observedRunningTime="2025-12-10 19:13:25.078487421 +0000 UTC m=+1087.873335213" watchObservedRunningTime="2025-12-10 19:13:25.087126416 +0000 UTC m=+1087.881974208" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.128447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" event={"ID":"6a960bb0-9f2c-4610-a418-791038fd1a06","Type":"ContainerStarted","Data":"8f3c3b15cf89432c17e0992f7b3308dcf163f102ecc4102ac8471689d796a57e"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.129443 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.137489 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.157950 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" event={"ID":"74c8428d-0015-4712-8bfa-6bf12f68b0bf","Type":"ContainerStarted","Data":"7676895d3f6bc31db559c418bdd222188f0325091de22f8b8a4697762dc7cbc2"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.158721 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.162618 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.163071 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" event={"ID":"08615ae3-02b0-41fb-8951-433da5904be5","Type":"ContainerStarted","Data":"1f2813ea7e02be8e131667f72652aac7329f27ea07e3e0051715513d25309fa1"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.163608 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6fvml" podStartSLOduration=4.292139812 podStartE2EDuration="33.163594896s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.872758756 +0000 UTC m=+1057.667606548" lastFinishedPulling="2025-12-10 19:13:23.74421384 +0000 UTC m=+1086.539061632" observedRunningTime="2025-12-10 19:13:25.16229215 +0000 UTC m=+1087.957139942" watchObservedRunningTime="2025-12-10 19:13:25.163594896 +0000 UTC m=+1087.958442688" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.183045 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" event={"ID":"7dd9f4f8-92dc-496a-a9b1-647a1619b7a5","Type":"ContainerStarted","Data":"24ca0c9f0edddf6b981394cff32e8aeb4e6ebc4a53a70befbb3a6ac13a25afc1"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.190791 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" event={"ID":"df0a10a8-f92d-4cd9-843c-8769455c9b24","Type":"ContainerStarted","Data":"5d3a905e9cf3ba3988ceaedbc59c440c1cd621083a7897b62748b3128e6e07a5"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.192804 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" podStartSLOduration=4.860395228 podStartE2EDuration="33.1927879s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:55.193059088 +0000 UTC m=+1057.987906870" lastFinishedPulling="2025-12-10 19:13:23.52545175 +0000 UTC m=+1086.320299542" observedRunningTime="2025-12-10 19:13:25.183571579 +0000 UTC m=+1087.978419371" watchObservedRunningTime="2025-12-10 19:13:25.1927879 +0000 UTC m=+1087.987635702" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.197545 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" event={"ID":"ea0450b1-401e-42af-9e33-f29b60af26ed","Type":"ContainerStarted","Data":"6515eef5c2ef47d3aa713009533029a12f51d52dbc7c4a835065255813a46211"} Dec 10 19:13:25 crc kubenswrapper[4894]: E1210 19:13:25.202197 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" podUID="ea0450b1-401e-42af-9e33-f29b60af26ed" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.215690 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.233492 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3bce49e-40c1-4e29-9905-825760e0f21d-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879frxskr\" (UID: \"b3bce49e-40c1-4e29-9905-825760e0f21d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.240364 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" event={"ID":"cc54976b-bc17-4cb5-8eb3-901a163c2ec8","Type":"ContainerStarted","Data":"cabca3dee00acd6aa18e375631a8a5a52dc393d8456110c4e39b49e477800c85"} Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.245294 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.246195 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-kkcfl" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.246390 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-jlzvg" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.332184 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xtknb" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.341304 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.372545 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rmkmh" podStartSLOduration=4.080195966 podStartE2EDuration="32.372521078s" podCreationTimestamp="2025-12-10 19:12:53 +0000 UTC" firstStartedPulling="2025-12-10 19:12:55.201410505 +0000 UTC m=+1057.996258297" lastFinishedPulling="2025-12-10 19:13:23.493735617 +0000 UTC m=+1086.288583409" observedRunningTime="2025-12-10 19:13:25.357296464 +0000 UTC m=+1088.152144276" watchObservedRunningTime="2025-12-10 19:13:25.372521078 +0000 UTC m=+1088.167368870" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.378191 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-xb72b" podStartSLOduration=4.163798629 podStartE2EDuration="33.378171732s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.269806625 +0000 UTC m=+1057.064654417" lastFinishedPulling="2025-12-10 19:13:23.484179728 +0000 UTC m=+1086.279027520" observedRunningTime="2025-12-10 19:13:25.377312569 +0000 UTC m=+1088.172160371" watchObservedRunningTime="2025-12-10 19:13:25.378171732 +0000 UTC m=+1088.173019534" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.448736 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b"] Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.550714 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" podStartSLOduration=3.794800775 podStartE2EDuration="33.550690445s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.888691589 +0000 UTC m=+1057.683539381" lastFinishedPulling="2025-12-10 19:13:24.644581259 +0000 UTC m=+1087.439429051" observedRunningTime="2025-12-10 19:13:25.524296727 +0000 UTC m=+1088.319144519" watchObservedRunningTime="2025-12-10 19:13:25.550690445 +0000 UTC m=+1088.345538237" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.948563 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.948625 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.956579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:25 crc kubenswrapper[4894]: I1210 19:13:25.958609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-x9dnx\" (UID: \"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.005382 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr"] Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.078693 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-st9lw" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.087462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.250957 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" event={"ID":"3a81ced4-fe81-41ca-9623-243b811b3491","Type":"ContainerStarted","Data":"f4cbe13ee75612a02a60837e1954a16b4fc2c7fbacfdb0aff5fd94d9c18d5fb2"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.251372 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.256477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" event={"ID":"08615ae3-02b0-41fb-8951-433da5904be5","Type":"ContainerStarted","Data":"642986def0b06cb173f314d899865295e75c4da8ce2dfca685f538b898a00e87"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.257047 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.261632 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" event={"ID":"7dd9f4f8-92dc-496a-a9b1-647a1619b7a5","Type":"ContainerStarted","Data":"e44ae4e55bcd8745af888f50bcf6e32b86dc45e70b5616169f6cb101e83ae414"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.262590 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.267753 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" event={"ID":"ddfceacd-628c-4f2f-bdc5-2c8cd7329098","Type":"ContainerStarted","Data":"639d7380b864b15a8c5bbd97123dafc2fa1f88b4b0c37eb8bcbfada6ff006c08"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.267859 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.284469 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" podStartSLOduration=3.449419501 podStartE2EDuration="34.284449373s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.08796747 +0000 UTC m=+1056.882815262" lastFinishedPulling="2025-12-10 19:13:24.922997342 +0000 UTC m=+1087.717845134" observedRunningTime="2025-12-10 19:13:26.268600871 +0000 UTC m=+1089.063448663" watchObservedRunningTime="2025-12-10 19:13:26.284449373 +0000 UTC m=+1089.079297165" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.295747 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" event={"ID":"119c37b8-34ea-403c-9d87-6f252ff9a983","Type":"ContainerStarted","Data":"5e857fb19ec9c54ac6a93e02bcd355315a8e03c0eaed4d11602407bdaadd0339"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.307650 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" podStartSLOduration=3.556617186 podStartE2EDuration="34.307625243s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.994806515 +0000 UTC m=+1057.789654307" lastFinishedPulling="2025-12-10 19:13:25.745814572 +0000 UTC m=+1088.540662364" observedRunningTime="2025-12-10 19:13:26.296042758 +0000 UTC m=+1089.090890550" watchObservedRunningTime="2025-12-10 19:13:26.307625243 +0000 UTC m=+1089.102473035" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.309183 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" event={"ID":"b3bce49e-40c1-4e29-9905-825760e0f21d","Type":"ContainerStarted","Data":"fc7ad7623b1dad176c024811a3bddcf29d30de61f93098e832df8124a2e27965"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.322798 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" podStartSLOduration=2.56042927 podStartE2EDuration="33.322780425s" podCreationTimestamp="2025-12-10 19:12:53 +0000 UTC" firstStartedPulling="2025-12-10 19:12:55.172363795 +0000 UTC m=+1057.967211587" lastFinishedPulling="2025-12-10 19:13:25.93471495 +0000 UTC m=+1088.729562742" observedRunningTime="2025-12-10 19:13:26.320307868 +0000 UTC m=+1089.115155660" watchObservedRunningTime="2025-12-10 19:13:26.322780425 +0000 UTC m=+1089.117628217" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.328443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" event={"ID":"badfd685-2701-4ced-a77e-6eb7b7ff8c01","Type":"ContainerStarted","Data":"65137abbd73cf9b9ebe48796e8e2fed587e911a81c176091efb455bbf379819a"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.328499 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" event={"ID":"df0a10a8-f92d-4cd9-843c-8769455c9b24","Type":"ContainerStarted","Data":"e67aef7fbba94d8ca826991927b1b335523502a752a3df688a1c4603320d04da"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.328515 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" event={"ID":"e1ef63d6-b70d-4424-86a3-988318cb4241","Type":"ContainerStarted","Data":"97cedc83aa761d684b61057fc88ea4f8b24d519e8d7e4fc61dd2b7c0c7ff5018"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.328537 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.328552 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.328567 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.330320 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" event={"ID":"1a65ee6c-f87e-4183-8c8c-9582347febda","Type":"ContainerStarted","Data":"d0aa38f97075670489823a634ab31e583e987837e856a9b8bcffa9ffd6f812a0"} Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.330355 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:13:26 crc kubenswrapper[4894]: E1210 19:13:26.332828 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" podUID="ea0450b1-401e-42af-9e33-f29b60af26ed" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.357211 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" podStartSLOduration=3.023762934 podStartE2EDuration="33.357192422s" podCreationTimestamp="2025-12-10 19:12:53 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.973311351 +0000 UTC m=+1057.768159143" lastFinishedPulling="2025-12-10 19:13:25.306740839 +0000 UTC m=+1088.101588631" observedRunningTime="2025-12-10 19:13:26.355409322 +0000 UTC m=+1089.150257134" watchObservedRunningTime="2025-12-10 19:13:26.357192422 +0000 UTC m=+1089.152040224" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.446007 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" podStartSLOduration=3.920516194 podStartE2EDuration="34.445991926s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.316191558 +0000 UTC m=+1057.111039350" lastFinishedPulling="2025-12-10 19:13:24.84166729 +0000 UTC m=+1087.636515082" observedRunningTime="2025-12-10 19:13:26.390871888 +0000 UTC m=+1089.185719690" watchObservedRunningTime="2025-12-10 19:13:26.445991926 +0000 UTC m=+1089.240839718" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.490449 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" podStartSLOduration=3.817279626 podStartE2EDuration="34.490433075s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.970117774 +0000 UTC m=+1057.764965566" lastFinishedPulling="2025-12-10 19:13:25.643271223 +0000 UTC m=+1088.438119015" observedRunningTime="2025-12-10 19:13:26.485801119 +0000 UTC m=+1089.280648911" watchObservedRunningTime="2025-12-10 19:13:26.490433075 +0000 UTC m=+1089.285280867" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.514800 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" podStartSLOduration=3.3204191610000002 podStartE2EDuration="34.514785397s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.751691413 +0000 UTC m=+1057.546539205" lastFinishedPulling="2025-12-10 19:13:25.946057649 +0000 UTC m=+1088.740905441" observedRunningTime="2025-12-10 19:13:26.511508359 +0000 UTC m=+1089.306356141" watchObservedRunningTime="2025-12-10 19:13:26.514785397 +0000 UTC m=+1089.309633189" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.556149 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" podStartSLOduration=3.153388128 podStartE2EDuration="34.556130772s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.532324816 +0000 UTC m=+1057.327172608" lastFinishedPulling="2025-12-10 19:13:25.93506746 +0000 UTC m=+1088.729915252" observedRunningTime="2025-12-10 19:13:26.552858733 +0000 UTC m=+1089.347706535" watchObservedRunningTime="2025-12-10 19:13:26.556130772 +0000 UTC m=+1089.350978564" Dec 10 19:13:26 crc kubenswrapper[4894]: I1210 19:13:26.608250 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx"] Dec 10 19:13:27 crc kubenswrapper[4894]: I1210 19:13:27.339324 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" event={"ID":"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd","Type":"ContainerStarted","Data":"4933c44944c123711874bfdb5de9967b69d17af4b28a2c3402b5ea892e2c75f9"} Dec 10 19:13:32 crc kubenswrapper[4894]: I1210 19:13:32.830445 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-99gdz" Dec 10 19:13:32 crc kubenswrapper[4894]: I1210 19:13:32.862791 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-c7qtj" Dec 10 19:13:32 crc kubenswrapper[4894]: I1210 19:13:32.930780 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lq68m" Dec 10 19:13:33 crc kubenswrapper[4894]: I1210 19:13:33.056194 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-6fk8l" Dec 10 19:13:33 crc kubenswrapper[4894]: I1210 19:13:33.285803 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4dvmk" Dec 10 19:13:33 crc kubenswrapper[4894]: I1210 19:13:33.592609 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-4dkpz" Dec 10 19:13:33 crc kubenswrapper[4894]: I1210 19:13:33.657146 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-qsjsk" Dec 10 19:13:33 crc kubenswrapper[4894]: I1210 19:13:33.680758 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-cg4zq" Dec 10 19:13:33 crc kubenswrapper[4894]: I1210 19:13:33.972502 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-c96bw" Dec 10 19:13:34 crc kubenswrapper[4894]: I1210 19:13:34.072001 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-m8wb7" Dec 10 19:13:34 crc kubenswrapper[4894]: I1210 19:13:34.088528 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-qlq67" Dec 10 19:13:36 crc kubenswrapper[4894]: I1210 19:13:36.432013 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" event={"ID":"5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd","Type":"ContainerStarted","Data":"6e49e3c7f62b03c7989136899fe684700ba895cf6a3529527ee0d79ff33e0b1e"} Dec 10 19:13:37 crc kubenswrapper[4894]: E1210 19:13:37.124519 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2223915556/1\": happened during read: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48" Dec 10 19:13:37 crc kubenswrapper[4894]: E1210 19:13:37.125864 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7nzl4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-84b575879frxskr_openstack-operators(b3bce49e-40c1-4e29-9905-825760e0f21d): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2223915556/1\": happened during read: context canceled" logger="UnhandledError" Dec 10 19:13:37 crc kubenswrapper[4894]: I1210 19:13:37.440830 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:39 crc kubenswrapper[4894]: E1210 19:13:39.962668 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage2223915556/1\\\": happened during read: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" podUID="b3bce49e-40c1-4e29-9905-825760e0f21d" Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.471223 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" event={"ID":"0f98c7f8-f73d-4ca7-8361-91a719ab74ae","Type":"ContainerStarted","Data":"48d87bebe2750e9b6d313512bdb25565b38ab20b4fde02e4489f086ff0d10dd9"} Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.471345 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.472691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" event={"ID":"119c37b8-34ea-403c-9d87-6f252ff9a983","Type":"ContainerStarted","Data":"902f8b84640ba5c61127d580ac6f1e19882eff9d091b76d42a24d0d4a8e8719c"} Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.472729 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" event={"ID":"119c37b8-34ea-403c-9d87-6f252ff9a983","Type":"ContainerStarted","Data":"4f2f2dc7816d4747ad551ae14a23c5d2641470effdd56739a801d1e3877c10eb"} Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.472793 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.473772 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" event={"ID":"b3bce49e-40c1-4e29-9905-825760e0f21d","Type":"ContainerStarted","Data":"78677c9ac6d65cbe5584f56ee407c9cc0a1c45396094f55d3b6d901a36ccd941"} Dec 10 19:13:40 crc kubenswrapper[4894]: E1210 19:13:40.475026 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" podUID="b3bce49e-40c1-4e29-9905-825760e0f21d" Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.489554 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" podStartSLOduration=3.720067292 podStartE2EDuration="48.48953179s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.994754964 +0000 UTC m=+1057.789602756" lastFinishedPulling="2025-12-10 19:13:39.764219462 +0000 UTC m=+1102.559067254" observedRunningTime="2025-12-10 19:13:40.487063443 +0000 UTC m=+1103.281911235" watchObservedRunningTime="2025-12-10 19:13:40.48953179 +0000 UTC m=+1103.284379582" Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.513305 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" podStartSLOduration=47.513280906 podStartE2EDuration="47.513280906s" podCreationTimestamp="2025-12-10 19:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:13:37.480570839 +0000 UTC m=+1100.275418671" watchObservedRunningTime="2025-12-10 19:13:40.513280906 +0000 UTC m=+1103.308128718" Dec 10 19:13:40 crc kubenswrapper[4894]: I1210 19:13:40.545352 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" podStartSLOduration=34.237430763 podStartE2EDuration="48.545335258s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:13:25.462897067 +0000 UTC m=+1088.257744849" lastFinishedPulling="2025-12-10 19:13:39.770801552 +0000 UTC m=+1102.565649344" observedRunningTime="2025-12-10 19:13:40.533857506 +0000 UTC m=+1103.328705308" watchObservedRunningTime="2025-12-10 19:13:40.545335258 +0000 UTC m=+1103.340183050" Dec 10 19:13:41 crc kubenswrapper[4894]: I1210 19:13:41.486511 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" event={"ID":"ea0450b1-401e-42af-9e33-f29b60af26ed","Type":"ContainerStarted","Data":"8bf4346e7acf0940be1c570d4f9421da5b8b43b9615b7e65449e500080abfdcc"} Dec 10 19:13:41 crc kubenswrapper[4894]: E1210 19:13:41.492091 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" podUID="b3bce49e-40c1-4e29-9905-825760e0f21d" Dec 10 19:13:41 crc kubenswrapper[4894]: I1210 19:13:41.507629 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" podStartSLOduration=3.333430965 podStartE2EDuration="49.507607311s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:12:54.850663614 +0000 UTC m=+1057.645511406" lastFinishedPulling="2025-12-10 19:13:41.02483996 +0000 UTC m=+1103.819687752" observedRunningTime="2025-12-10 19:13:41.506347427 +0000 UTC m=+1104.301195229" watchObservedRunningTime="2025-12-10 19:13:41.507607311 +0000 UTC m=+1104.302455113" Dec 10 19:13:43 crc kubenswrapper[4894]: I1210 19:13:43.080469 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:13:44 crc kubenswrapper[4894]: I1210 19:13:44.848873 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-wqv7b" Dec 10 19:13:46 crc kubenswrapper[4894]: I1210 19:13:46.095110 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-x9dnx" Dec 10 19:13:53 crc kubenswrapper[4894]: I1210 19:13:53.084623 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wphkl" Dec 10 19:13:53 crc kubenswrapper[4894]: I1210 19:13:53.527817 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2j6cg" Dec 10 19:13:55 crc kubenswrapper[4894]: I1210 19:13:55.626571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" event={"ID":"b3bce49e-40c1-4e29-9905-825760e0f21d","Type":"ContainerStarted","Data":"a3e63f635a7e5516f293443a7f0f67c774677e2826580fa235ad700433adde20"} Dec 10 19:13:55 crc kubenswrapper[4894]: I1210 19:13:55.626927 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:13:55 crc kubenswrapper[4894]: I1210 19:13:55.666494 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" podStartSLOduration=35.000339231 podStartE2EDuration="1m3.666473078s" podCreationTimestamp="2025-12-10 19:12:52 +0000 UTC" firstStartedPulling="2025-12-10 19:13:26.028967503 +0000 UTC m=+1088.823815295" lastFinishedPulling="2025-12-10 19:13:54.69510135 +0000 UTC m=+1117.489949142" observedRunningTime="2025-12-10 19:13:55.660908976 +0000 UTC m=+1118.455756778" watchObservedRunningTime="2025-12-10 19:13:55.666473078 +0000 UTC m=+1118.461320890" Dec 10 19:14:05 crc kubenswrapper[4894]: I1210 19:14:05.349935 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879frxskr" Dec 10 19:14:19 crc kubenswrapper[4894]: I1210 19:14:19.577356 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:14:19 crc kubenswrapper[4894]: I1210 19:14:19.577886 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.229882 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vd8cs"] Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.237636 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.241618 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wcltm" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.242148 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.253281 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.260258 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.297613 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vd8cs"] Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.315839 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ss55s"] Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.317089 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.321229 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.329324 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ss55s"] Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.432929 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msjj2\" (UniqueName: \"kubernetes.io/projected/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-kube-api-access-msjj2\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.432971 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753178fa-fec5-4360-ac26-ebb886b468e2-config\") pod \"dnsmasq-dns-675f4bcbfc-vd8cs\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.432996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42tkv\" (UniqueName: \"kubernetes.io/projected/753178fa-fec5-4360-ac26-ebb886b468e2-kube-api-access-42tkv\") pod \"dnsmasq-dns-675f4bcbfc-vd8cs\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.433016 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.433501 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-config\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.534897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-config\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.534954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msjj2\" (UniqueName: \"kubernetes.io/projected/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-kube-api-access-msjj2\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.534973 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753178fa-fec5-4360-ac26-ebb886b468e2-config\") pod \"dnsmasq-dns-675f4bcbfc-vd8cs\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.534992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42tkv\" (UniqueName: \"kubernetes.io/projected/753178fa-fec5-4360-ac26-ebb886b468e2-kube-api-access-42tkv\") pod \"dnsmasq-dns-675f4bcbfc-vd8cs\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.535014 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.535689 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.535745 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-config\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.536015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753178fa-fec5-4360-ac26-ebb886b468e2-config\") pod \"dnsmasq-dns-675f4bcbfc-vd8cs\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.554951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42tkv\" (UniqueName: \"kubernetes.io/projected/753178fa-fec5-4360-ac26-ebb886b468e2-kube-api-access-42tkv\") pod \"dnsmasq-dns-675f4bcbfc-vd8cs\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.557307 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msjj2\" (UniqueName: \"kubernetes.io/projected/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-kube-api-access-msjj2\") pod \"dnsmasq-dns-78dd6ddcc-ss55s\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.595001 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:21 crc kubenswrapper[4894]: I1210 19:14:21.640084 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:22 crc kubenswrapper[4894]: I1210 19:14:22.048186 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vd8cs"] Dec 10 19:14:22 crc kubenswrapper[4894]: I1210 19:14:22.115895 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ss55s"] Dec 10 19:14:22 crc kubenswrapper[4894]: W1210 19:14:22.119408 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod635f3d8e_bef2_4b9a_a583_a03f2d0b1edb.slice/crio-69c9ece237468835b2cc2a0eda71f49d681bb020646ea314ab051b9a9e4430e2 WatchSource:0}: Error finding container 69c9ece237468835b2cc2a0eda71f49d681bb020646ea314ab051b9a9e4430e2: Status 404 returned error can't find the container with id 69c9ece237468835b2cc2a0eda71f49d681bb020646ea314ab051b9a9e4430e2 Dec 10 19:14:22 crc kubenswrapper[4894]: I1210 19:14:22.882621 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" event={"ID":"753178fa-fec5-4360-ac26-ebb886b468e2","Type":"ContainerStarted","Data":"a893c03dda4b45126d47ea5434d45b6c752088e96ebe6326baa0f8af674266ad"} Dec 10 19:14:22 crc kubenswrapper[4894]: I1210 19:14:22.883590 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" event={"ID":"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb","Type":"ContainerStarted","Data":"69c9ece237468835b2cc2a0eda71f49d681bb020646ea314ab051b9a9e4430e2"} Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.288014 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vd8cs"] Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.318374 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tstn2"] Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.334424 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.346464 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tstn2"] Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.483665 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gnwn\" (UniqueName: \"kubernetes.io/projected/c41065a9-9e5a-4157-b094-6524bc953df1-kube-api-access-6gnwn\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.483759 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-config\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.483792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.587159 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gnwn\" (UniqueName: \"kubernetes.io/projected/c41065a9-9e5a-4157-b094-6524bc953df1-kube-api-access-6gnwn\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.587208 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-config\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.587233 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.589600 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-config\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.589947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.595647 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ss55s"] Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.630306 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gnwn\" (UniqueName: \"kubernetes.io/projected/c41065a9-9e5a-4157-b094-6524bc953df1-kube-api-access-6gnwn\") pod \"dnsmasq-dns-666b6646f7-tstn2\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.646875 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jnlkn"] Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.648257 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.653756 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jnlkn"] Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.677207 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.789959 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfwdb\" (UniqueName: \"kubernetes.io/projected/e6752b66-a898-4b95-a065-01c0da2024ef-kube-api-access-nfwdb\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.790044 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-config\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.790210 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.891219 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.891305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfwdb\" (UniqueName: \"kubernetes.io/projected/e6752b66-a898-4b95-a065-01c0da2024ef-kube-api-access-nfwdb\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.891345 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-config\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.892181 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-config\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.894412 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.947232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfwdb\" (UniqueName: \"kubernetes.io/projected/e6752b66-a898-4b95-a065-01c0da2024ef-kube-api-access-nfwdb\") pod \"dnsmasq-dns-57d769cc4f-jnlkn\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:24 crc kubenswrapper[4894]: I1210 19:14:24.971207 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.422296 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tstn2"] Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.484158 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.492478 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.494422 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.494719 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.494757 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.495019 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.495150 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7xk4b" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.495403 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.495630 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.496493 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.556142 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jnlkn"] Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605315 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kb2m\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-kube-api-access-2kb2m\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605369 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605432 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97355ddc-532e-46a2-a568-31f4a5e07030-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605465 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605547 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605570 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605585 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97355ddc-532e-46a2-a568-31f4a5e07030-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605776 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.605875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-config-data\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708261 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97355ddc-532e-46a2-a568-31f4a5e07030-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708312 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708340 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-config-data\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708380 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kb2m\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-kube-api-access-2kb2m\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708458 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708479 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708508 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97355ddc-532e-46a2-a568-31f4a5e07030-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708549 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.708790 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.709000 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.710093 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-config-data\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.710467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.710502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.713101 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.713130 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.713178 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c975b249a0e4a5d819d74290b52a518dd30d8c0b64914f36ad533f66100f6cb7/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.716703 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97355ddc-532e-46a2-a568-31f4a5e07030-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.720683 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.732254 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97355ddc-532e-46a2-a568-31f4a5e07030-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.733357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kb2m\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-kube-api-access-2kb2m\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.751322 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.774789 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.779547 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.786947 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.787628 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.789236 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.789403 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.789468 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.789622 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.789880 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wnk5r" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.790028 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.823298 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911129 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911207 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911235 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911253 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911281 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911332 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911350 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911368 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2zbf\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-kube-api-access-d2zbf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.911410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.930882 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" event={"ID":"e6752b66-a898-4b95-a065-01c0da2024ef","Type":"ContainerStarted","Data":"1e680624e91378c63bf9ab7a638162748640bd4a41b939bb08772dd42341f4e8"} Dec 10 19:14:25 crc kubenswrapper[4894]: I1210 19:14:25.933171 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" event={"ID":"c41065a9-9e5a-4157-b094-6524bc953df1","Type":"ContainerStarted","Data":"9abe5a81331b6ba03b3de64d88f4c839cb230acc83bc099df35c4067b5dbb870"} Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014559 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014607 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014639 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2zbf\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-kube-api-access-d2zbf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014654 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014726 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014904 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.014966 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.015018 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.015073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.015799 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.017627 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.017784 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.018410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.019507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.020406 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.020559 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.021021 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.024990 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.025018 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8fde555e75f624d7793c9908446e92e60787eecb4e949c1f5f68bb115b06429c/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.030345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.034260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2zbf\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-kube-api-access-d2zbf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.059148 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.104017 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.295403 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.684609 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:14:26 crc kubenswrapper[4894]: W1210 19:14:26.699661 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f81cda0_aeed_41c1_9941_5f1ba92a88c5.slice/crio-bb7c046121b938465f7c400dcff686e5d628a64787e9a627274c2bf17d2e9ea3 WatchSource:0}: Error finding container bb7c046121b938465f7c400dcff686e5d628a64787e9a627274c2bf17d2e9ea3: Status 404 returned error can't find the container with id bb7c046121b938465f7c400dcff686e5d628a64787e9a627274c2bf17d2e9ea3 Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.942534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8f81cda0-aeed-41c1-9941-5f1ba92a88c5","Type":"ContainerStarted","Data":"bb7c046121b938465f7c400dcff686e5d628a64787e9a627274c2bf17d2e9ea3"} Dec 10 19:14:26 crc kubenswrapper[4894]: I1210 19:14:26.946200 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97355ddc-532e-46a2-a568-31f4a5e07030","Type":"ContainerStarted","Data":"d17bbf1ddf0f3043f97b69700cdf0d4b10ed61ce020314125ab026833e5a2c1f"} Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.203214 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.204510 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.206365 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.208056 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-89mkg" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.208199 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.209309 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.212709 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.219272 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.366755 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0297c834-eee7-4414-ba18-b4ee6b95561f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367219 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-config-data-default\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367250 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0297c834-eee7-4414-ba18-b4ee6b95561f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367315 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367353 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0297c834-eee7-4414-ba18-b4ee6b95561f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367403 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftgsm\" (UniqueName: \"kubernetes.io/projected/0297c834-eee7-4414-ba18-b4ee6b95561f-kube-api-access-ftgsm\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367443 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.367477 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-kolla-config\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.474896 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-kolla-config\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.474942 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0297c834-eee7-4414-ba18-b4ee6b95561f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.474971 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-config-data-default\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.474997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0297c834-eee7-4414-ba18-b4ee6b95561f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.475056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.475072 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0297c834-eee7-4414-ba18-b4ee6b95561f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.475100 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftgsm\" (UniqueName: \"kubernetes.io/projected/0297c834-eee7-4414-ba18-b4ee6b95561f-kube-api-access-ftgsm\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.475143 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.476545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.476980 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-kolla-config\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.479655 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0297c834-eee7-4414-ba18-b4ee6b95561f-config-data-default\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.481035 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0297c834-eee7-4414-ba18-b4ee6b95561f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.490813 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0297c834-eee7-4414-ba18-b4ee6b95561f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.491737 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.491764 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a179e0f725a53c5073786ceae2c26d9178aa92e82500c518229657e84bd0f0d4/globalmount\"" pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.493189 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0297c834-eee7-4414-ba18-b4ee6b95561f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.502824 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftgsm\" (UniqueName: \"kubernetes.io/projected/0297c834-eee7-4414-ba18-b4ee6b95561f-kube-api-access-ftgsm\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.546074 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de0aee46-8f4e-4e31-bb95-eaf24b6631f0\") pod \"openstack-galera-0\" (UID: \"0297c834-eee7-4414-ba18-b4ee6b95561f\") " pod="openstack/openstack-galera-0" Dec 10 19:14:27 crc kubenswrapper[4894]: I1210 19:14:27.832236 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.524764 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.526605 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.529593 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.529619 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.530678 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.531757 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-t89q5" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.534126 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692256 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldst7\" (UniqueName: \"kubernetes.io/projected/ab75b963-c89e-4871-8cb1-180dbaaf2714-kube-api-access-ldst7\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab75b963-c89e-4871-8cb1-180dbaaf2714-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692380 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab75b963-c89e-4871-8cb1-180dbaaf2714-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692416 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692436 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab75b963-c89e-4871-8cb1-180dbaaf2714-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692535 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.692568 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794488 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab75b963-c89e-4871-8cb1-180dbaaf2714-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794599 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794656 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab75b963-c89e-4871-8cb1-180dbaaf2714-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794732 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794809 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldst7\" (UniqueName: \"kubernetes.io/projected/ab75b963-c89e-4871-8cb1-180dbaaf2714-kube-api-access-ldst7\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab75b963-c89e-4871-8cb1-180dbaaf2714-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794956 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab75b963-c89e-4871-8cb1-180dbaaf2714-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.794980 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.798229 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.798252 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.799233 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.799258 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e3f66eded730c53e5974f5ae38820fbc0fa3b90cf94312da433a42a429bf2249/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.804186 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab75b963-c89e-4871-8cb1-180dbaaf2714-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.805209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab75b963-c89e-4871-8cb1-180dbaaf2714-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.814503 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldst7\" (UniqueName: \"kubernetes.io/projected/ab75b963-c89e-4871-8cb1-180dbaaf2714-kube-api-access-ldst7\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.829928 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab75b963-c89e-4871-8cb1-180dbaaf2714-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.881247 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-31fa626f-7795-4a2a-94df-7d51d4b0e0b0\") pod \"openstack-cell1-galera-0\" (UID: \"ab75b963-c89e-4871-8cb1-180dbaaf2714\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.917259 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.921352 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.921451 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.924812 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-g6dxj" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.924986 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 10 19:14:28 crc kubenswrapper[4894]: I1210 19:14:28.925004 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.111292 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7f34d67-8b26-449d-b38f-524e352efdd2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.111345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f34d67-8b26-449d-b38f-524e352efdd2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.111377 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7f34d67-8b26-449d-b38f-524e352efdd2-kolla-config\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.111403 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpfrw\" (UniqueName: \"kubernetes.io/projected/c7f34d67-8b26-449d-b38f-524e352efdd2-kube-api-access-wpfrw\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.111442 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7f34d67-8b26-449d-b38f-524e352efdd2-config-data\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.150221 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.213095 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7f34d67-8b26-449d-b38f-524e352efdd2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.213131 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f34d67-8b26-449d-b38f-524e352efdd2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.213158 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7f34d67-8b26-449d-b38f-524e352efdd2-kolla-config\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.213178 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpfrw\" (UniqueName: \"kubernetes.io/projected/c7f34d67-8b26-449d-b38f-524e352efdd2-kube-api-access-wpfrw\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.213470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7f34d67-8b26-449d-b38f-524e352efdd2-config-data\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.214083 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7f34d67-8b26-449d-b38f-524e352efdd2-kolla-config\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.214360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c7f34d67-8b26-449d-b38f-524e352efdd2-config-data\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.219309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7f34d67-8b26-449d-b38f-524e352efdd2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.229534 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f34d67-8b26-449d-b38f-524e352efdd2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.232372 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpfrw\" (UniqueName: \"kubernetes.io/projected/c7f34d67-8b26-449d-b38f-524e352efdd2-kube-api-access-wpfrw\") pod \"memcached-0\" (UID: \"c7f34d67-8b26-449d-b38f-524e352efdd2\") " pod="openstack/memcached-0" Dec 10 19:14:29 crc kubenswrapper[4894]: I1210 19:14:29.274943 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.673838 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.675123 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.679225 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gz9zs" Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.694514 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.838679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8pt\" (UniqueName: \"kubernetes.io/projected/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d-kube-api-access-bf8pt\") pod \"kube-state-metrics-0\" (UID: \"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d\") " pod="openstack/kube-state-metrics-0" Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.940496 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8pt\" (UniqueName: \"kubernetes.io/projected/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d-kube-api-access-bf8pt\") pod \"kube-state-metrics-0\" (UID: \"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d\") " pod="openstack/kube-state-metrics-0" Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.968392 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8pt\" (UniqueName: \"kubernetes.io/projected/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d-kube-api-access-bf8pt\") pod \"kube-state-metrics-0\" (UID: \"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d\") " pod="openstack/kube-state-metrics-0" Dec 10 19:14:30 crc kubenswrapper[4894]: I1210 19:14:30.996254 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.359748 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.361973 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.363579 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.363740 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.363908 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.364038 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-5npqs" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.364333 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.385513 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.549444 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88520de0-eedd-40d7-ad33-ee752e8f2ff5-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.549528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqxcm\" (UniqueName: \"kubernetes.io/projected/88520de0-eedd-40d7-ad33-ee752e8f2ff5-kube-api-access-jqxcm\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.549595 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88520de0-eedd-40d7-ad33-ee752e8f2ff5-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.549641 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.550322 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.550352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/88520de0-eedd-40d7-ad33-ee752e8f2ff5-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.550378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652123 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88520de0-eedd-40d7-ad33-ee752e8f2ff5-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652174 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqxcm\" (UniqueName: \"kubernetes.io/projected/88520de0-eedd-40d7-ad33-ee752e8f2ff5-kube-api-access-jqxcm\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88520de0-eedd-40d7-ad33-ee752e8f2ff5-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652264 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652287 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/88520de0-eedd-40d7-ad33-ee752e8f2ff5-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.652713 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/88520de0-eedd-40d7-ad33-ee752e8f2ff5-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.656705 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.658368 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.660349 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/88520de0-eedd-40d7-ad33-ee752e8f2ff5-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.660634 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/88520de0-eedd-40d7-ad33-ee752e8f2ff5-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.662345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/88520de0-eedd-40d7-ad33-ee752e8f2ff5-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.674380 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqxcm\" (UniqueName: \"kubernetes.io/projected/88520de0-eedd-40d7-ad33-ee752e8f2ff5-kube-api-access-jqxcm\") pod \"alertmanager-metric-storage-0\" (UID: \"88520de0-eedd-40d7-ad33-ee752e8f2ff5\") " pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.720556 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.988233 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.990045 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.994795 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-d22s9" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.995012 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.995120 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.995236 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 10 19:14:31 crc kubenswrapper[4894]: I1210 19:14:31.995998 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.020747 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.028794 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.158899 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159133 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t2lj\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-kube-api-access-6t2lj\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159166 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159325 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f3a6b99-5cd8-4149-8825-e19613a13135-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159449 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f3a6b99-5cd8-4149-8825-e19613a13135-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.159473 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261379 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t2lj\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-kube-api-access-6t2lj\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f3a6b99-5cd8-4149-8825-e19613a13135-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261452 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f3a6b99-5cd8-4149-8825-e19613a13135-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.261511 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.262871 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f3a6b99-5cd8-4149-8825-e19613a13135-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.265642 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.266579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.268086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.270160 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f3a6b99-5cd8-4149-8825-e19613a13135-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.290486 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.290528 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ee4633b447c807486101aee2fd0379a07b7444d2812f770a7ff8debe13b85c97/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.293976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.308603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t2lj\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-kube-api-access-6t2lj\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.376502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:32 crc kubenswrapper[4894]: I1210 19:14:32.627471 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.512987 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-h2552"] Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.514463 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.517421 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-fwmsv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.517624 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.517792 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.573006 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h2552"] Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.581426 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-pdwjv"] Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.584028 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.587409 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pdwjv"] Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-combined-ca-bundle\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629573 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-run-ovn\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629634 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-scripts\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629662 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pf6m\" (UniqueName: \"kubernetes.io/projected/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-kube-api-access-2pf6m\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629704 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-run\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629757 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-log-ovn\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.629783 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-ovn-controller-tls-certs\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.730873 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-etc-ovs\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.730912 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-run\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.730941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-combined-ca-bundle\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731111 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h9wc\" (UniqueName: \"kubernetes.io/projected/df6eb3cc-342e-4d18-8027-db2c751fbaab-kube-api-access-9h9wc\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731237 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-run-ovn\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731270 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-lib\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731302 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df6eb3cc-342e-4d18-8027-db2c751fbaab-scripts\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731543 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-scripts\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pf6m\" (UniqueName: \"kubernetes.io/projected/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-kube-api-access-2pf6m\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-run\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731718 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-log\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731783 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-log-ovn\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731815 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-ovn-controller-tls-certs\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.731911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-run-ovn\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.732260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-run\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.732639 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-var-log-ovn\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.733817 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-scripts\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.736163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-ovn-controller-tls-certs\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.736817 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-combined-ca-bundle\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.751715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pf6m\" (UniqueName: \"kubernetes.io/projected/5fd5a07b-caf3-44d6-a9d4-e50e5be1819c-kube-api-access-2pf6m\") pod \"ovn-controller-h2552\" (UID: \"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c\") " pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.833685 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-etc-ovs\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.833992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-run\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834108 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-run\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h9wc\" (UniqueName: \"kubernetes.io/projected/df6eb3cc-342e-4d18-8027-db2c751fbaab-kube-api-access-9h9wc\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834235 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-etc-ovs\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-lib\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df6eb3cc-342e-4d18-8027-db2c751fbaab-scripts\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834588 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-log\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834719 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-lib\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.834804 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/df6eb3cc-342e-4d18-8027-db2c751fbaab-var-log\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.838079 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df6eb3cc-342e-4d18-8027-db2c751fbaab-scripts\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.839010 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.852319 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h9wc\" (UniqueName: \"kubernetes.io/projected/df6eb3cc-342e-4d18-8027-db2c751fbaab-kube-api-access-9h9wc\") pod \"ovn-controller-ovs-pdwjv\" (UID: \"df6eb3cc-342e-4d18-8027-db2c751fbaab\") " pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:35 crc kubenswrapper[4894]: I1210 19:14:35.906885 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.414969 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.416319 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.423065 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.423141 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.423353 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.423512 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tlhwz" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.423651 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.439460 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580328 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580396 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580418 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcg2s\" (UniqueName: \"kubernetes.io/projected/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-kube-api-access-bcg2s\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580442 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580472 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580520 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-config\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.580540 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.681902 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.681970 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.681996 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-config\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.682028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.682116 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.682164 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.682194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcg2s\" (UniqueName: \"kubernetes.io/projected/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-kube-api-access-bcg2s\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.682225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.682402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.684327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-config\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.684524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.686811 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.687143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.688117 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.688176 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cd2b0015f5a92dd507bbef8ef9067831c851265b9844d471095099a2c988dc21/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.688565 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.717434 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2462fd9-632d-4ace-94fb-40ffe6a43d40\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.718742 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcg2s\" (UniqueName: \"kubernetes.io/projected/f7ec72a7-3043-4044-adfc-7ca2bf0c1749-kube-api-access-bcg2s\") pod \"ovsdbserver-nb-0\" (UID: \"f7ec72a7-3043-4044-adfc-7ca2bf0c1749\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:36 crc kubenswrapper[4894]: I1210 19:14:36.801552 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.635187 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.636747 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.639007 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.639042 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-ftcd8" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.639067 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.639047 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.649005 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.814803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.814895 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.814944 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.814988 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.815145 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5thsj\" (UniqueName: \"kubernetes.io/projected/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-kube-api-access-5thsj\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.815190 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.815269 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.815345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-config\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.916969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917041 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917088 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5thsj\" (UniqueName: \"kubernetes.io/projected/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-kube-api-access-5thsj\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917113 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917154 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-config\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917253 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.917744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.918292 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.918329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-config\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.922479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.922885 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.922923 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/999db3375bc79bea3b8e5e28a7c425438d3d18a62b8deae471f80da0540dcfcd/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.924456 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.925588 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.942278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5thsj\" (UniqueName: \"kubernetes.io/projected/5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b-kube-api-access-5thsj\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.962905 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1ecaf19a-3594-4552-9274-9a52e6b9b327\") pod \"ovsdbserver-sb-0\" (UID: \"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:38 crc kubenswrapper[4894]: I1210 19:14:38.966176 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.617545 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65"] Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.619226 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.623410 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.628485 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.628737 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-q6xk4" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.628884 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.628934 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.669320 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65"] Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.703888 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.703926 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.703969 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6bdq\" (UniqueName: \"kubernetes.io/projected/028272b8-eb6a-4981-86b0-9438be5c55d3-kube-api-access-v6bdq\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.703997 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.704061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028272b8-eb6a-4981-86b0-9438be5c55d3-config\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.805938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.805982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.806022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6bdq\" (UniqueName: \"kubernetes.io/projected/028272b8-eb6a-4981-86b0-9438be5c55d3-kube-api-access-v6bdq\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.806048 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.806124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028272b8-eb6a-4981-86b0-9438be5c55d3-config\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.806609 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x"] Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.807117 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028272b8-eb6a-4981-86b0-9438be5c55d3-config\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.807605 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.808671 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.811098 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.811194 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-loki-s3" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.811403 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.813019 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.819449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/028272b8-eb6a-4981-86b0-9438be5c55d3-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.822264 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x"] Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.856481 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6bdq\" (UniqueName: \"kubernetes.io/projected/028272b8-eb6a-4981-86b0-9438be5c55d3-kube-api-access-v6bdq\") pod \"cloudkitty-lokistack-distributor-66dfd9bb-4qx65\" (UID: \"028272b8-eb6a-4981-86b0-9438be5c55d3\") " pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.883103 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h"] Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.884189 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.888578 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.888813 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.894346 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h"] Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.907275 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tnmh\" (UniqueName: \"kubernetes.io/projected/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-kube-api-access-7tnmh\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.907686 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.907722 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.907810 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.907917 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-config\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.907945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.939140 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:14:43 crc kubenswrapper[4894]: I1210 19:14:43.999835 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.006518 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.008941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-config\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009045 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009132 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tnmh\" (UniqueName: \"kubernetes.io/projected/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-kube-api-access-7tnmh\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009418 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009514 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldjpt\" (UniqueName: \"kubernetes.io/projected/09e53cf2-d31c-4270-9002-29d42f94ea00-kube-api-access-ldjpt\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009596 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009669 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009761 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e53cf2-d31c-4270-9002-29d42f94ea00-config\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.009847 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.017544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-config\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.018174 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.018467 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-ppz9c" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.018661 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.018839 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.019923 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.021529 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.022251 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.024461 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.019147 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.019230 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.019290 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.020036 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.028549 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.028647 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.047921 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.054558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tnmh\" (UniqueName: \"kubernetes.io/projected/bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae-kube-api-access-7tnmh\") pod \"cloudkitty-lokistack-querier-795fd8f8cc-rhw5x\" (UID: \"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae\") " pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.112957 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113002 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113059 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-rbac\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113081 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-tenants\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113109 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-tenants\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113141 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldjpt\" (UniqueName: \"kubernetes.io/projected/09e53cf2-d31c-4270-9002-29d42f94ea00-kube-api-access-ldjpt\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113305 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113456 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113500 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-tls-secret\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113540 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113560 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e53cf2-d31c-4270-9002-29d42f94ea00-config\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113643 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113668 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-tls-secret\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113700 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkpk5\" (UniqueName: \"kubernetes.io/projected/e3561537-bc07-4e04-8e80-7ff3b376d8c9-kube-api-access-fkpk5\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113716 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg6wx\" (UniqueName: \"kubernetes.io/projected/9f05b972-ebf9-4b08-95c0-224225ef4271-kube-api-access-bg6wx\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.113736 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-rbac\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.114631 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.114799 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09e53cf2-d31c-4270-9002-29d42f94ea00-config\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.124528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.125869 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/09e53cf2-d31c-4270-9002-29d42f94ea00-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.128603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldjpt\" (UniqueName: \"kubernetes.io/projected/09e53cf2-d31c-4270-9002-29d42f94ea00-kube-api-access-ldjpt\") pod \"cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h\" (UID: \"09e53cf2-d31c-4270-9002-29d42f94ea00\") " pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.214764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215277 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-rbac\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215358 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-tenants\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215437 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-tenants\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215888 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215974 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216049 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-tls-secret\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216180 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-rbac\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216388 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216416 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216435 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216467 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-tls-secret\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216526 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkpk5\" (UniqueName: \"kubernetes.io/projected/e3561537-bc07-4e04-8e80-7ff3b376d8c9-kube-api-access-fkpk5\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216548 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg6wx\" (UniqueName: \"kubernetes.io/projected/9f05b972-ebf9-4b08-95c0-224225ef4271-kube-api-access-bg6wx\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.216574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-rbac\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.215883 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.217047 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.217108 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.217168 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.217598 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-rbac\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.219558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.219758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-tenants\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.220341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-tls-secret\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.220602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.221720 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/e3561537-bc07-4e04-8e80-7ff3b376d8c9-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.223449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-tenants\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.224835 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/9f05b972-ebf9-4b08-95c0-224225ef4271-tls-secret\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.237738 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg6wx\" (UniqueName: \"kubernetes.io/projected/9f05b972-ebf9-4b08-95c0-224225ef4271-kube-api-access-bg6wx\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl\" (UID: \"9f05b972-ebf9-4b08-95c0-224225ef4271\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.238447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkpk5\" (UniqueName: \"kubernetes.io/projected/e3561537-bc07-4e04-8e80-7ff3b376d8c9-kube-api-access-fkpk5\") pod \"cloudkitty-lokistack-gateway-7db4f4db8c-75nc2\" (UID: \"e3561537-bc07-4e04-8e80-7ff3b376d8c9\") " pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.239073 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.249674 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.384548 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.397213 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.779898 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.781130 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.783309 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.784253 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.797509 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.855715 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.857132 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.859634 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.860746 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.871614 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.945287 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.946409 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.947954 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.947991 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948046 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948067 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948126 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948160 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldbgs\" (UniqueName: \"kubernetes.io/projected/da36895a-2989-4110-88a2-6bf0eb5d2f5d-kube-api-access-ldbgs\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948192 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bf8748-8765-4f93-8edf-c4755dfe89bb-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948263 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klj88\" (UniqueName: \"kubernetes.io/projected/25bf8748-8765-4f93-8edf-c4755dfe89bb-kube-api-access-klj88\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948311 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948438 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da36895a-2989-4110-88a2-6bf0eb5d2f5d-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948490 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.948521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.949150 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.949328 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Dec 10 19:14:44 crc kubenswrapper[4894]: I1210 19:14:44.985547 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.050140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.050463 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.050560 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdlmh\" (UniqueName: \"kubernetes.io/projected/d5987e81-6ef3-4d95-88dc-7c05d531075f-kube-api-access-qdlmh\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.050657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.050736 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.050856 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051005 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051434 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051560 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051670 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldbgs\" (UniqueName: \"kubernetes.io/projected/da36895a-2989-4110-88a2-6bf0eb5d2f5d-kube-api-access-ldbgs\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051745 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bf8748-8765-4f93-8edf-c4755dfe89bb-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051813 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051828 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.051970 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5987e81-6ef3-4d95-88dc-7c05d531075f-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052089 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klj88\" (UniqueName: \"kubernetes.io/projected/25bf8748-8765-4f93-8edf-c4755dfe89bb-kube-api-access-klj88\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052154 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da36895a-2989-4110-88a2-6bf0eb5d2f5d-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052214 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.052810 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.054053 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.055065 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bf8748-8765-4f93-8edf-c4755dfe89bb-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.055135 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da36895a-2989-4110-88a2-6bf0eb5d2f5d-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.055640 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.055940 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.056449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.056968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.057369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.059700 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.061689 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/25bf8748-8765-4f93-8edf-c4755dfe89bb-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.065067 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/da36895a-2989-4110-88a2-6bf0eb5d2f5d-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.071413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klj88\" (UniqueName: \"kubernetes.io/projected/25bf8748-8765-4f93-8edf-c4755dfe89bb-kube-api-access-klj88\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.088629 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldbgs\" (UniqueName: \"kubernetes.io/projected/da36895a-2989-4110-88a2-6bf0eb5d2f5d-kube-api-access-ldbgs\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.096958 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.110261 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"25bf8748-8765-4f93-8edf-c4755dfe89bb\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.120827 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"da36895a-2989-4110-88a2-6bf0eb5d2f5d\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154172 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5987e81-6ef3-4d95-88dc-7c05d531075f-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154263 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154314 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154335 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdlmh\" (UniqueName: \"kubernetes.io/projected/d5987e81-6ef3-4d95-88dc-7c05d531075f-kube-api-access-qdlmh\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154367 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.154422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.155139 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.155346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5987e81-6ef3-4d95-88dc-7c05d531075f-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.155491 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.159365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.159861 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.163320 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/d5987e81-6ef3-4d95-88dc-7c05d531075f-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.172781 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdlmh\" (UniqueName: \"kubernetes.io/projected/d5987e81-6ef3-4d95-88dc-7c05d531075f-kube-api-access-qdlmh\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.181130 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"d5987e81-6ef3-4d95-88dc-7c05d531075f\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.181186 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.255268 4894 patch_prober.go:28] interesting pod/router-default-5444994796-mk6q8 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.255332 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-mk6q8" podUID="37a3cb31-fac7-4cc6-aa07-e5db7be4dba4" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.278938 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:14:45 crc kubenswrapper[4894]: I1210 19:14:45.397609 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:14:48 crc kubenswrapper[4894]: E1210 19:14:48.436418 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 10 19:14:48 crc kubenswrapper[4894]: E1210 19:14:48.436896 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2kb2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(97355ddc-532e-46a2-a568-31f4a5e07030): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:14:48 crc kubenswrapper[4894]: E1210 19:14:48.438256 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" Dec 10 19:14:48 crc kubenswrapper[4894]: E1210 19:14:48.483096 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 10 19:14:48 crc kubenswrapper[4894]: E1210 19:14:48.483340 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d2zbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(8f81cda0-aeed-41c1-9941-5f1ba92a88c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:14:48 crc kubenswrapper[4894]: E1210 19:14:48.485345 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.128716 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.129227 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nfwdb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-jnlkn_openstack(e6752b66-a898-4b95-a065-01c0da2024ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.130926 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.140670 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.140834 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gnwn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-tstn2_openstack(c41065a9-9e5a-4157-b094-6524bc953df1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.142021 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.165186 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.165211 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.169668 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.246937 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.247073 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-42tkv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-vd8cs_openstack(753178fa-fec5-4360-ac26-ebb886b468e2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.248158 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" podUID="753178fa-fec5-4360-ac26-ebb886b468e2" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.262188 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.262380 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msjj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-ss55s_openstack(635f3d8e-bef2-4b9a-a583-a03f2d0b1edb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:14:49 crc kubenswrapper[4894]: E1210 19:14:49.263763 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" podUID="635f3d8e-bef2-4b9a-a583-a03f2d0b1edb" Dec 10 19:14:49 crc kubenswrapper[4894]: I1210 19:14:49.576982 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:14:49 crc kubenswrapper[4894]: I1210 19:14:49.577264 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:14:49 crc kubenswrapper[4894]: I1210 19:14:49.943213 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.021091 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.159749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88520de0-eedd-40d7-ad33-ee752e8f2ff5","Type":"ContainerStarted","Data":"3ce6bf6a29b345632bafe4ddc79db4bd06b8c9816b29dc15ba1741ff49616c99"} Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.160487 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0297c834-eee7-4414-ba18-b4ee6b95561f","Type":"ContainerStarted","Data":"de0f02c6d739b515f16b10db8eba34e107be1dcba342dcb75e45363d669dfd45"} Dec 10 19:14:50 crc kubenswrapper[4894]: E1210 19:14:50.161767 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.580340 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.599099 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:14:50 crc kubenswrapper[4894]: W1210 19:14:50.602988 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod028272b8_eb6a_4981_86b0_9438be5c55d3.slice/crio-7a69549945168e0ed84fc9617c8ed46785b4ac0cd82733744c643fd71299055d WatchSource:0}: Error finding container 7a69549945168e0ed84fc9617c8ed46785b4ac0cd82733744c643fd71299055d: Status 404 returned error can't find the container with id 7a69549945168e0ed84fc9617c8ed46785b4ac0cd82733744c643fd71299055d Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.622187 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.639022 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.652603 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.665567 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42tkv\" (UniqueName: \"kubernetes.io/projected/753178fa-fec5-4360-ac26-ebb886b468e2-kube-api-access-42tkv\") pod \"753178fa-fec5-4360-ac26-ebb886b468e2\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.665614 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753178fa-fec5-4360-ac26-ebb886b468e2-config\") pod \"753178fa-fec5-4360-ac26-ebb886b468e2\" (UID: \"753178fa-fec5-4360-ac26-ebb886b468e2\") " Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.666696 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/753178fa-fec5-4360-ac26-ebb886b468e2-config" (OuterVolumeSpecName: "config") pod "753178fa-fec5-4360-ac26-ebb886b468e2" (UID: "753178fa-fec5-4360-ac26-ebb886b468e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:14:50 crc kubenswrapper[4894]: W1210 19:14:50.666767 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab75b963_c89e_4871_8cb1_180dbaaf2714.slice/crio-fc3d47d39ba1931ab6a423b120cd32a2e9bf2ab675d01dee3a6844c2c3b89fdf WatchSource:0}: Error finding container fc3d47d39ba1931ab6a423b120cd32a2e9bf2ab675d01dee3a6844c2c3b89fdf: Status 404 returned error can't find the container with id fc3d47d39ba1931ab6a423b120cd32a2e9bf2ab675d01dee3a6844c2c3b89fdf Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.671405 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.679913 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/753178fa-fec5-4360-ac26-ebb886b468e2-kube-api-access-42tkv" (OuterVolumeSpecName: "kube-api-access-42tkv") pod "753178fa-fec5-4360-ac26-ebb886b468e2" (UID: "753178fa-fec5-4360-ac26-ebb886b468e2"). InnerVolumeSpecName "kube-api-access-42tkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.683182 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.754488 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.761153 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h2552"] Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.766760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-dns-svc\") pod \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.766811 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msjj2\" (UniqueName: \"kubernetes.io/projected/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-kube-api-access-msjj2\") pod \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.767022 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-config\") pod \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\" (UID: \"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb\") " Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.767455 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42tkv\" (UniqueName: \"kubernetes.io/projected/753178fa-fec5-4360-ac26-ebb886b468e2-kube-api-access-42tkv\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.767474 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753178fa-fec5-4360-ac26-ebb886b468e2-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.768207 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-config" (OuterVolumeSpecName: "config") pod "635f3d8e-bef2-4b9a-a583-a03f2d0b1edb" (UID: "635f3d8e-bef2-4b9a-a583-a03f2d0b1edb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.768669 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "635f3d8e-bef2-4b9a-a583-a03f2d0b1edb" (UID: "635f3d8e-bef2-4b9a-a583-a03f2d0b1edb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.775114 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-kube-api-access-msjj2" (OuterVolumeSpecName: "kube-api-access-msjj2") pod "635f3d8e-bef2-4b9a-a583-a03f2d0b1edb" (UID: "635f3d8e-bef2-4b9a-a583-a03f2d0b1edb"). InnerVolumeSpecName "kube-api-access-msjj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.869104 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.869143 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msjj2\" (UniqueName: \"kubernetes.io/projected/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-kube-api-access-msjj2\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.869156 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:50 crc kubenswrapper[4894]: I1210 19:14:50.919832 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pdwjv"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.018675 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.028515 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl"] Dec 10 19:14:51 crc kubenswrapper[4894]: W1210 19:14:51.035701 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f05b972_ebf9_4b08_95c0_224225ef4271.slice/crio-2528bb1b2dd787456b8d8fdd40f4cebdbba6979846d1cd4fcc87467c6450a440 WatchSource:0}: Error finding container 2528bb1b2dd787456b8d8fdd40f4cebdbba6979846d1cd4fcc87467c6450a440: Status 404 returned error can't find the container with id 2528bb1b2dd787456b8d8fdd40f4cebdbba6979846d1cd4fcc87467c6450a440 Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.037461 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 19:14:51 crc kubenswrapper[4894]: W1210 19:14:51.038035 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb4a57d8_a817_4bc8_9ffa_7f151d44d9ae.slice/crio-0bfba2b9844f1034a21ab2a6940292f8da95d8d302e14934f9fb30428491214a WatchSource:0}: Error finding container 0bfba2b9844f1034a21ab2a6940292f8da95d8d302e14934f9fb30428491214a: Status 404 returned error can't find the container with id 0bfba2b9844f1034a21ab2a6940292f8da95d8d302e14934f9fb30428491214a Dec 10 19:14:51 crc kubenswrapper[4894]: W1210 19:14:51.039209 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7f34d67_8b26_449d_b38f_524e352efdd2.slice/crio-263edb3d6f8336227c1c4e0028cefab3032e0aab6802827c4ff459fd70b491dc WatchSource:0}: Error finding container 263edb3d6f8336227c1c4e0028cefab3032e0aab6802827c4ff459fd70b491dc: Status 404 returned error can't find the container with id 263edb3d6f8336227c1c4e0028cefab3032e0aab6802827c4ff459fd70b491dc Dec 10 19:14:51 crc kubenswrapper[4894]: W1210 19:14:51.044590 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3561537_bc07_4e04_8e80_7ff3b376d8c9.slice/crio-deed1ab8eee823f49fbf99d7f92250cb3b1e07251107b517650551ede639b63e WatchSource:0}: Error finding container deed1ab8eee823f49fbf99d7f92250cb3b1e07251107b517650551ede639b63e: Status 404 returned error can't find the container with id deed1ab8eee823f49fbf99d7f92250cb3b1e07251107b517650551ede639b63e Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.047823 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.057877 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.064936 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 10 19:14:51 crc kubenswrapper[4894]: W1210 19:14:51.069740 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda36895a_2989_4110_88a2_6bf0eb5d2f5d.slice/crio-a3412c87baa3f946be570c78e7d0856cc5cb0e09f725ed167ce04bc4cb366ea4 WatchSource:0}: Error finding container a3412c87baa3f946be570c78e7d0856cc5cb0e09f725ed167ce04bc4cb366ea4: Status 404 returned error can't find the container with id a3412c87baa3f946be570c78e7d0856cc5cb0e09f725ed167ce04bc4cb366ea4 Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.071466 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-ingester,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d,Command:[],Args:[-target=ingester -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:gossip-ring,HostPort:0,ContainerPort:7946,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:wal,ReadOnly:false,MountPath:/tmp/wal,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ingester-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ingester-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ldbgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-ingester-0_openstack(da36895a-2989-4110-88a2-6bf0eb5d2f5d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.072387 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-compactor,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d,Command:[],Args:[-target=compactor -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-compactor-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-compactor-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-klj88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-compactor-0_openstack(25bf8748-8765-4f93-8edf-c4755dfe89bb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.073017 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.073228 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-ingester\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="da36895a-2989-4110-88a2-6bf0eb5d2f5d" Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.073578 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="25bf8748-8765-4f93-8edf-c4755dfe89bb" Dec 10 19:14:51 crc kubenswrapper[4894]: W1210 19:14:51.073676 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5987e81_6ef3_4d95_88dc_7c05d531075f.slice/crio-c956d31bbf1be82090c8b2c1f29baecb3712a474bf6c198218c7ce7de4e7f164 WatchSource:0}: Error finding container c956d31bbf1be82090c8b2c1f29baecb3712a474bf6c198218c7ce7de4e7f164: Status 404 returned error can't find the container with id c956d31bbf1be82090c8b2c1f29baecb3712a474bf6c198218c7ce7de4e7f164 Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.078303 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-index-gateway,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d,Command:[],Args:[-target=index-gateway -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:storage,ReadOnly:false,MountPath:/tmp/loki,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-index-gateway-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-index-gateway-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qdlmh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-index-gateway-0_openstack(d5987e81-6ef3-4d95-88dc-7c05d531075f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.080022 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-index-gateway\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" podUID="d5987e81-6ef3-4d95-88dc-7c05d531075f" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.103422 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.108977 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n547h567h9ch5dbh8dh54dh5cchb9h5cdh687h9dh569hbch54ch6fhcfhdch558h5h87h67dh8hf7hfdh568h595hd5hbch658hc9h57h6q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bcg2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(f7ec72a7-3043-4044-adfc-7ca2bf0c1749): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.111071 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n547h567h9ch5dbh8dh54dh5cchb9h5cdh687h9dh569hbch54ch6fhcfhdch558h5h87h67dh8hf7hfdh568h595hd5hbch658hc9h57h6q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bcg2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(f7ec72a7-3043-4044-adfc-7ca2bf0c1749): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.112236 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack/ovsdbserver-nb-0" podUID="f7ec72a7-3043-4044-adfc-7ca2bf0c1749" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.180522 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" event={"ID":"028272b8-eb6a-4981-86b0-9438be5c55d3","Type":"ContainerStarted","Data":"7a69549945168e0ed84fc9617c8ed46785b4ac0cd82733744c643fd71299055d"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.181707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" event={"ID":"9f05b972-ebf9-4b08-95c0-224225ef4271","Type":"ContainerStarted","Data":"2528bb1b2dd787456b8d8fdd40f4cebdbba6979846d1cd4fcc87467c6450a440"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.182634 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"d5987e81-6ef3-4d95-88dc-7c05d531075f","Type":"ContainerStarted","Data":"c956d31bbf1be82090c8b2c1f29baecb3712a474bf6c198218c7ce7de4e7f164"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.183981 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab75b963-c89e-4871-8cb1-180dbaaf2714","Type":"ContainerStarted","Data":"fc3d47d39ba1931ab6a423b120cd32a2e9bf2ab675d01dee3a6844c2c3b89fdf"} Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.184043 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-index-gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" podUID="d5987e81-6ef3-4d95-88dc-7c05d531075f" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.186162 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" event={"ID":"09e53cf2-d31c-4270-9002-29d42f94ea00","Type":"ContainerStarted","Data":"39ad4e895fa0d085edc1b3fad439396538b9cc09feed6af0fc144eff4ad2004c"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.187614 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"da36895a-2989-4110-88a2-6bf0eb5d2f5d","Type":"ContainerStarted","Data":"a3412c87baa3f946be570c78e7d0856cc5cb0e09f725ed167ce04bc4cb366ea4"} Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.189254 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-ingester\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="da36895a-2989-4110-88a2-6bf0eb5d2f5d" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.190163 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b","Type":"ContainerStarted","Data":"0dacd1d35bfba04fadb348db20d8f479f4d73f1664a1394d08fd4e850d94f360"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.193033 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" event={"ID":"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae","Type":"ContainerStarted","Data":"0bfba2b9844f1034a21ab2a6940292f8da95d8d302e14934f9fb30428491214a"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.195004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" event={"ID":"753178fa-fec5-4360-ac26-ebb886b468e2","Type":"ContainerDied","Data":"a893c03dda4b45126d47ea5434d45b6c752088e96ebe6326baa0f8af674266ad"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.195188 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-vd8cs" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.212144 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552" event={"ID":"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c","Type":"ContainerStarted","Data":"b97b82b81a07c5427f35d14051b9992c3bd7c6304868c09f58bf2346cf01d0bb"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.214507 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"25bf8748-8765-4f93-8edf-c4755dfe89bb","Type":"ContainerStarted","Data":"96355480fdaf64f89458a308db1ae19362713e61be6a4d24c5213904fb5be83e"} Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.216186 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="25bf8748-8765-4f93-8edf-c4755dfe89bb" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.217157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerStarted","Data":"5edc5921c54b5bf44ceb2939bf610e5b1771551145fd39870376bd643a6bcf32"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.218341 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d","Type":"ContainerStarted","Data":"fb9d70af946d96dc2705db5b2cd8cd2c46828855bcb074ec498e1d5434e271a8"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.219361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f7ec72a7-3043-4044-adfc-7ca2bf0c1749","Type":"ContainerStarted","Data":"09da7795f652d8fb1db5dd145d9fab0d7c9bffd95a6232bb76618db2677fe203"} Dec 10 19:14:51 crc kubenswrapper[4894]: E1210 19:14:51.224671 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\", failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"]" pod="openstack/ovsdbserver-nb-0" podUID="f7ec72a7-3043-4044-adfc-7ca2bf0c1749" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.233702 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c7f34d67-8b26-449d-b38f-524e352efdd2","Type":"ContainerStarted","Data":"263edb3d6f8336227c1c4e0028cefab3032e0aab6802827c4ff459fd70b491dc"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.236215 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" event={"ID":"e3561537-bc07-4e04-8e80-7ff3b376d8c9","Type":"ContainerStarted","Data":"deed1ab8eee823f49fbf99d7f92250cb3b1e07251107b517650551ede639b63e"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.237565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" event={"ID":"635f3d8e-bef2-4b9a-a583-a03f2d0b1edb","Type":"ContainerDied","Data":"69c9ece237468835b2cc2a0eda71f49d681bb020646ea314ab051b9a9e4430e2"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.237632 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ss55s" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.242277 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pdwjv" event={"ID":"df6eb3cc-342e-4d18-8027-db2c751fbaab","Type":"ContainerStarted","Data":"0ec46b99a762de0ec6ebdde13da7934fef0cde3e665314f6a344d2899949defc"} Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.302836 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vd8cs"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.323989 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-vd8cs"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.393173 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ss55s"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.399187 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ss55s"] Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.554253 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="635f3d8e-bef2-4b9a-a583-a03f2d0b1edb" path="/var/lib/kubelet/pods/635f3d8e-bef2-4b9a-a583-a03f2d0b1edb/volumes" Dec 10 19:14:51 crc kubenswrapper[4894]: I1210 19:14:51.554980 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="753178fa-fec5-4360-ac26-ebb886b468e2" path="/var/lib/kubelet/pods/753178fa-fec5-4360-ac26-ebb886b468e2/volumes" Dec 10 19:14:52 crc kubenswrapper[4894]: E1210 19:14:52.252974 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-index-gateway\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" podUID="d5987e81-6ef3-4d95-88dc-7c05d531075f" Dec 10 19:14:52 crc kubenswrapper[4894]: E1210 19:14:52.254248 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-ingester\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="da36895a-2989-4110-88a2-6bf0eb5d2f5d" Dec 10 19:14:52 crc kubenswrapper[4894]: E1210 19:14:52.254392 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-compactor\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-compactor-0" podUID="25bf8748-8765-4f93-8edf-c4755dfe89bb" Dec 10 19:14:52 crc kubenswrapper[4894]: E1210 19:14:52.255515 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\", failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"]" pod="openstack/ovsdbserver-nb-0" podUID="f7ec72a7-3043-4044-adfc-7ca2bf0c1749" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.137015 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn"] Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.138725 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.140649 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.141432 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.145196 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn"] Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.262455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b191d21-d1ab-4df8-9e29-6747e121a047-secret-volume\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.262811 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgp2\" (UniqueName: \"kubernetes.io/projected/7b191d21-d1ab-4df8-9e29-6747e121a047-kube-api-access-czgp2\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.262942 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b191d21-d1ab-4df8-9e29-6747e121a047-config-volume\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.364272 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b191d21-d1ab-4df8-9e29-6747e121a047-config-volume\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.364357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b191d21-d1ab-4df8-9e29-6747e121a047-secret-volume\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.364387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czgp2\" (UniqueName: \"kubernetes.io/projected/7b191d21-d1ab-4df8-9e29-6747e121a047-kube-api-access-czgp2\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.365266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b191d21-d1ab-4df8-9e29-6747e121a047-config-volume\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.377500 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b191d21-d1ab-4df8-9e29-6747e121a047-secret-volume\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.403322 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgp2\" (UniqueName: \"kubernetes.io/projected/7b191d21-d1ab-4df8-9e29-6747e121a047-kube-api-access-czgp2\") pod \"collect-profiles-29423235-zm2dn\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:00 crc kubenswrapper[4894]: I1210 19:15:00.472839 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.312681 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.313627 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ftgsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(0297c834-eee7-4414-ba18-b4ee6b95561f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.314948 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="0297c834-eee7-4414-ba18-b4ee6b95561f" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.345622 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.346015 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ldst7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(ab75b963-c89e-4871-8cb1-180dbaaf2714): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.347263 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="ab75b963-c89e-4871-8cb1-180dbaaf2714" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.348232 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.348497 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:loki-querier,Image:registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d,Command:[],Args:[-target=querier -config.file=/etc/loki/config/config.yaml -runtime-config.file=/etc/loki/config/runtime-config.yaml -config.expand-env=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:3100,Protocol:TCP,HostIP:,},ContainerPort{Name:grpclb,HostPort:0,ContainerPort:9095,Protocol:TCP,HostIP:,},ContainerPort{Name:gossip-ring,HostPort:0,ContainerPort:7946,Protocol:TCP,HostIP:,},ContainerPort{Name:healthchecks,HostPort:0,ContainerPort:3101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:AWS_ACCESS_KEY_ID,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_id,Optional:nil,},},},EnvVar{Name:AWS_ACCESS_KEY_SECRET,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:cloudkitty-loki-s3,},Key:access_key_secret,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:false,MountPath:/etc/loki/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-querier-http,ReadOnly:false,MountPath:/var/run/tls/http/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-loki-s3,ReadOnly:false,MountPath:/etc/storage/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-querier-grpc,ReadOnly:false,MountPath:/var/run/tls/grpc/server,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cloudkitty-lokistack-ca-bundle,ReadOnly:false,MountPath:/var/run/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7tnmh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/loki/api/v1/status/buildinfo,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:2,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/ready,Port:{0 3101 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-lokistack-querier-795fd8f8cc-rhw5x_openstack(bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.349812 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-querier\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" podUID="bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.398081 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="ab75b963-c89e-4871-8cb1-180dbaaf2714" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.398549 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="0297c834-eee7-4414-ba18-b4ee6b95561f" Dec 10 19:15:09 crc kubenswrapper[4894]: E1210 19:15:09.398660 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"loki-querier\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:2b491fcb180423632d30811515a439a7a7f41023c1cfe4780647f18969b85a1d\\\"\"" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" podUID="bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae" Dec 10 19:15:10 crc kubenswrapper[4894]: I1210 19:15:10.721948 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn"] Dec 10 19:15:11 crc kubenswrapper[4894]: W1210 19:15:11.191142 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b191d21_d1ab_4df8_9e29_6747e121a047.slice/crio-7951ee9fcd7c0425a1c79749e4a61e7ad1518e62aa1fe63854d7a6c362d59d80 WatchSource:0}: Error finding container 7951ee9fcd7c0425a1c79749e4a61e7ad1518e62aa1fe63854d7a6c362d59d80: Status 404 returned error can't find the container with id 7951ee9fcd7c0425a1c79749e4a61e7ad1518e62aa1fe63854d7a6c362d59d80 Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.423580 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" event={"ID":"9f05b972-ebf9-4b08-95c0-224225ef4271","Type":"ContainerStarted","Data":"4e2d40d3b7ec4cae02a2d3b014dff01a690a69893a5a2af0fd104c406e5deb77"} Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.423777 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.427748 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" event={"ID":"09e53cf2-d31c-4270-9002-29d42f94ea00","Type":"ContainerStarted","Data":"6ba7dfb3dde777478b5ca88b6b5c843f5813d984abcd8fbf23d3cd6629eeabc4"} Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.428325 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.431877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" event={"ID":"028272b8-eb6a-4981-86b0-9438be5c55d3","Type":"ContainerStarted","Data":"c5ba01673007a54915ab0fdac58944ba26215a0256e6e2c559678b726b0dd812"} Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.432559 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.434941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" event={"ID":"7b191d21-d1ab-4df8-9e29-6747e121a047","Type":"ContainerStarted","Data":"7951ee9fcd7c0425a1c79749e4a61e7ad1518e62aa1fe63854d7a6c362d59d80"} Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.437153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c7f34d67-8b26-449d-b38f-524e352efdd2","Type":"ContainerStarted","Data":"d2a42e7d62b52e093414bdce010ad61df593241636333847338024cfcc61cd0a"} Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.437332 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.450344 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" podStartSLOduration=9.752501581 podStartE2EDuration="28.450319706s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.038669453 +0000 UTC m=+1173.833517245" lastFinishedPulling="2025-12-10 19:15:09.736487558 +0000 UTC m=+1192.531335370" observedRunningTime="2025-12-10 19:15:11.448834105 +0000 UTC m=+1194.243681907" watchObservedRunningTime="2025-12-10 19:15:11.450319706 +0000 UTC m=+1194.245167518" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.456281 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.478402 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" podStartSLOduration=9.260234638 podStartE2EDuration="28.478377369s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.654349847 +0000 UTC m=+1173.449197639" lastFinishedPulling="2025-12-10 19:15:09.872492568 +0000 UTC m=+1192.667340370" observedRunningTime="2025-12-10 19:15:11.470019322 +0000 UTC m=+1194.264867124" watchObservedRunningTime="2025-12-10 19:15:11.478377369 +0000 UTC m=+1194.273225161" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.491840 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" podStartSLOduration=8.930768424 podStartE2EDuration="28.491818715s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.614052551 +0000 UTC m=+1173.408900343" lastFinishedPulling="2025-12-10 19:15:10.175102842 +0000 UTC m=+1192.969950634" observedRunningTime="2025-12-10 19:15:11.488198386 +0000 UTC m=+1194.283046188" watchObservedRunningTime="2025-12-10 19:15:11.491818715 +0000 UTC m=+1194.286666527" Dec 10 19:15:11 crc kubenswrapper[4894]: I1210 19:15:11.510918 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=24.653358384 podStartE2EDuration="43.510898255s" podCreationTimestamp="2025-12-10 19:14:28 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.047783501 +0000 UTC m=+1173.842631293" lastFinishedPulling="2025-12-10 19:15:09.905323372 +0000 UTC m=+1192.700171164" observedRunningTime="2025-12-10 19:15:11.504400278 +0000 UTC m=+1194.299248070" watchObservedRunningTime="2025-12-10 19:15:11.510898255 +0000 UTC m=+1194.305746057" Dec 10 19:15:11 crc kubenswrapper[4894]: E1210 19:15:11.952173 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 10 19:15:11 crc kubenswrapper[4894]: E1210 19:15:11.952470 4894 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 10 19:15:11 crc kubenswrapper[4894]: E1210 19:15:11.952598 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bf8pt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(5fc7a32e-b6e9-4d8c-b007-b7aca357a98d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 19:15:11 crc kubenswrapper[4894]: E1210 19:15:11.953834 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" Dec 10 19:15:12 crc kubenswrapper[4894]: I1210 19:15:12.448503 4894 generic.go:334] "Generic (PLEG): container finished" podID="df6eb3cc-342e-4d18-8027-db2c751fbaab" containerID="ed5e157bc60328e64c2f5bac6d35f03079cf3568406edfbbf476d9080a7cd59e" exitCode=0 Dec 10 19:15:12 crc kubenswrapper[4894]: I1210 19:15:12.448637 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pdwjv" event={"ID":"df6eb3cc-342e-4d18-8027-db2c751fbaab","Type":"ContainerDied","Data":"ed5e157bc60328e64c2f5bac6d35f03079cf3568406edfbbf476d9080a7cd59e"} Dec 10 19:15:12 crc kubenswrapper[4894]: E1210 19:15:12.452721 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" Dec 10 19:15:13 crc kubenswrapper[4894]: I1210 19:15:13.462237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88520de0-eedd-40d7-ad33-ee752e8f2ff5","Type":"ContainerStarted","Data":"cf3ec44702652906c442f0b5ba02fc68571f55ed12e5c341d2b7a1a812232f57"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.471370 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f7ec72a7-3043-4044-adfc-7ca2bf0c1749","Type":"ContainerStarted","Data":"ea995be0675ca7e26c8408fb55d4cd62a4287ac92edff1cf7566fd8f664e93b5"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.474089 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"d5987e81-6ef3-4d95-88dc-7c05d531075f","Type":"ContainerStarted","Data":"3428975560d71374eb64200ad10146d6d7c1f5d042b9cff04ff8a1c337cc4aa6"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.474385 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.475861 4894 generic.go:334] "Generic (PLEG): container finished" podID="e6752b66-a898-4b95-a065-01c0da2024ef" containerID="0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344" exitCode=0 Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.475942 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" event={"ID":"e6752b66-a898-4b95-a065-01c0da2024ef","Type":"ContainerDied","Data":"0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.477776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" event={"ID":"e3561537-bc07-4e04-8e80-7ff3b376d8c9","Type":"ContainerStarted","Data":"b3063b2e0b44613d5d5bd4979f3510bfa9d4b671111dfc1a09e747f6a7403795"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.477902 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.480204 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" podUID="e3561537-bc07-4e04-8e80-7ff3b376d8c9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.122:8081/ready\": dial tcp 10.217.0.122:8081: connect: connection refused" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.480526 4894 generic.go:334] "Generic (PLEG): container finished" podID="c41065a9-9e5a-4157-b094-6524bc953df1" containerID="effbb40c043be21dc23390924b40b151fa8e72f242b74fa45a456da83010b865" exitCode=0 Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.480613 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" event={"ID":"c41065a9-9e5a-4157-b094-6524bc953df1","Type":"ContainerDied","Data":"effbb40c043be21dc23390924b40b151fa8e72f242b74fa45a456da83010b865"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.483053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"da36895a-2989-4110-88a2-6bf0eb5d2f5d","Type":"ContainerStarted","Data":"77720769d86eeba0cb90166a86881b6601883e62485c7990bbf7fed557597fb9"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.483355 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.485754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b","Type":"ContainerStarted","Data":"422ccbf996c468dc4477e24dc6c2f7045a15b60220d4bb097aca47d7052674e2"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.487659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552" event={"ID":"5fd5a07b-caf3-44d6-a9d4-e50e5be1819c","Type":"ContainerStarted","Data":"1c5240be6bbf9e71a0570f02faf1aaa6ff69ad66e9178c38feb063add2deb2f6"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.488438 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-h2552" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.491103 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" event={"ID":"7b191d21-d1ab-4df8-9e29-6747e121a047","Type":"ContainerStarted","Data":"f5e8428cc2584c191634b4c72457c154f95fc3661e66f4c4c45a958eb0f22ff4"} Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.508409 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=12.166902677 podStartE2EDuration="31.508365805s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.078144216 +0000 UTC m=+1173.872992008" lastFinishedPulling="2025-12-10 19:15:10.419607354 +0000 UTC m=+1193.214455136" observedRunningTime="2025-12-10 19:15:14.499419522 +0000 UTC m=+1197.294267334" watchObservedRunningTime="2025-12-10 19:15:14.508365805 +0000 UTC m=+1197.303213607" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.519061 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-h2552" podStartSLOduration=20.627038227 podStartE2EDuration="39.519038186s" podCreationTimestamp="2025-12-10 19:14:35 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.763773094 +0000 UTC m=+1173.558620886" lastFinishedPulling="2025-12-10 19:15:09.655773053 +0000 UTC m=+1192.450620845" observedRunningTime="2025-12-10 19:15:14.515956502 +0000 UTC m=+1197.310804324" watchObservedRunningTime="2025-12-10 19:15:14.519038186 +0000 UTC m=+1197.313885998" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.554053 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" podStartSLOduration=14.554032838 podStartE2EDuration="14.554032838s" podCreationTimestamp="2025-12-10 19:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:14.550686597 +0000 UTC m=+1197.345534389" watchObservedRunningTime="2025-12-10 19:15:14.554032838 +0000 UTC m=+1197.348880640" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.586964 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" podStartSLOduration=12.480603254 podStartE2EDuration="31.586933004s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.068168425 +0000 UTC m=+1173.863016217" lastFinishedPulling="2025-12-10 19:15:10.174498165 +0000 UTC m=+1192.969345967" observedRunningTime="2025-12-10 19:15:14.580601111 +0000 UTC m=+1197.375448913" watchObservedRunningTime="2025-12-10 19:15:14.586933004 +0000 UTC m=+1197.381780846" Dec 10 19:15:14 crc kubenswrapper[4894]: I1210 19:15:14.608892 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=12.282778631 podStartE2EDuration="31.60887122s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.07129269 +0000 UTC m=+1173.866140482" lastFinishedPulling="2025-12-10 19:15:10.397385279 +0000 UTC m=+1193.192233071" observedRunningTime="2025-12-10 19:15:14.602892397 +0000 UTC m=+1197.397740199" watchObservedRunningTime="2025-12-10 19:15:14.60887122 +0000 UTC m=+1197.403719032" Dec 10 19:15:14 crc kubenswrapper[4894]: E1210 19:15:14.835542 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6752b66_a898_4b95_a065_01c0da2024ef.slice/crio-conmon-0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b191d21_d1ab_4df8_9e29_6747e121a047.slice/crio-f5e8428cc2584c191634b4c72457c154f95fc3661e66f4c4c45a958eb0f22ff4.scope\": RecentStats: unable to find data in memory cache]" Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.505352 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pdwjv" event={"ID":"df6eb3cc-342e-4d18-8027-db2c751fbaab","Type":"ContainerStarted","Data":"9485b3556122b6ad1de5751024a5c406908a7f737b5ef1f9b7d4ae71d17eaff8"} Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.510225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" event={"ID":"e6752b66-a898-4b95-a065-01c0da2024ef","Type":"ContainerStarted","Data":"dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a"} Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.510892 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.512761 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"25bf8748-8765-4f93-8edf-c4755dfe89bb","Type":"ContainerStarted","Data":"366c404310fe968f2b3acceab113437406cb4956794852a32544f521a2c15040"} Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.512949 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.516041 4894 generic.go:334] "Generic (PLEG): container finished" podID="7b191d21-d1ab-4df8-9e29-6747e121a047" containerID="f5e8428cc2584c191634b4c72457c154f95fc3661e66f4c4c45a958eb0f22ff4" exitCode=0 Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.516153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" event={"ID":"7b191d21-d1ab-4df8-9e29-6747e121a047","Type":"ContainerDied","Data":"f5e8428cc2584c191634b4c72457c154f95fc3661e66f4c4c45a958eb0f22ff4"} Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.517622 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8f81cda0-aeed-41c1-9941-5f1ba92a88c5","Type":"ContainerStarted","Data":"f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e"} Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.519237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97355ddc-532e-46a2-a568-31f4a5e07030","Type":"ContainerStarted","Data":"cf236dedc39a8f67620519bcb9ba4e9cef45cc34cf202fb40082f8055745524a"} Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.537210 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-7db4f4db8c-75nc2" Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.540129 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" podStartSLOduration=6.586553737 podStartE2EDuration="51.540106716s" podCreationTimestamp="2025-12-10 19:14:24 +0000 UTC" firstStartedPulling="2025-12-10 19:14:25.563315561 +0000 UTC m=+1148.358163353" lastFinishedPulling="2025-12-10 19:15:10.51686854 +0000 UTC m=+1193.311716332" observedRunningTime="2025-12-10 19:15:15.531466811 +0000 UTC m=+1198.326314603" watchObservedRunningTime="2025-12-10 19:15:15.540106716 +0000 UTC m=+1198.334954508" Dec 10 19:15:15 crc kubenswrapper[4894]: I1210 19:15:15.599421 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=13.251056955 podStartE2EDuration="32.599401589s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.071892797 +0000 UTC m=+1173.866740589" lastFinishedPulling="2025-12-10 19:15:10.420237431 +0000 UTC m=+1193.215085223" observedRunningTime="2025-12-10 19:15:15.596585083 +0000 UTC m=+1198.391432885" watchObservedRunningTime="2025-12-10 19:15:15.599401589 +0000 UTC m=+1198.394249391" Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.546326 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pdwjv" event={"ID":"df6eb3cc-342e-4d18-8027-db2c751fbaab","Type":"ContainerStarted","Data":"fa931b1e2ef893dbd66ac226800c811990b3fd7d6b094772704cc98ef7bd4220"} Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.546658 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.551108 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerStarted","Data":"08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f"} Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.554438 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" event={"ID":"c41065a9-9e5a-4157-b094-6524bc953df1","Type":"ContainerStarted","Data":"3ff514733815170b51661556993c6e3340f6b6fa15fc1716859f14ea177cf54b"} Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.554784 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.590252 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" podStartSLOduration=7.626975133 podStartE2EDuration="52.590233386s" podCreationTimestamp="2025-12-10 19:14:24 +0000 UTC" firstStartedPulling="2025-12-10 19:14:25.434050554 +0000 UTC m=+1148.228898386" lastFinishedPulling="2025-12-10 19:15:10.397308837 +0000 UTC m=+1193.192156639" observedRunningTime="2025-12-10 19:15:16.583499974 +0000 UTC m=+1199.378347766" watchObservedRunningTime="2025-12-10 19:15:16.590233386 +0000 UTC m=+1199.385081178" Dec 10 19:15:16 crc kubenswrapper[4894]: I1210 19:15:16.590912 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-pdwjv" podStartSLOduration=22.870050473 podStartE2EDuration="41.590907105s" podCreationTimestamp="2025-12-10 19:14:35 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.934973582 +0000 UTC m=+1173.729821374" lastFinishedPulling="2025-12-10 19:15:09.655830214 +0000 UTC m=+1192.450678006" observedRunningTime="2025-12-10 19:15:16.568639599 +0000 UTC m=+1199.363487401" watchObservedRunningTime="2025-12-10 19:15:16.590907105 +0000 UTC m=+1199.385754887" Dec 10 19:15:17 crc kubenswrapper[4894]: I1210 19:15:17.560990 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.279432 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.405688 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b191d21-d1ab-4df8-9e29-6747e121a047-secret-volume\") pod \"7b191d21-d1ab-4df8-9e29-6747e121a047\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.405770 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b191d21-d1ab-4df8-9e29-6747e121a047-config-volume\") pod \"7b191d21-d1ab-4df8-9e29-6747e121a047\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.406559 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b191d21-d1ab-4df8-9e29-6747e121a047-config-volume" (OuterVolumeSpecName: "config-volume") pod "7b191d21-d1ab-4df8-9e29-6747e121a047" (UID: "7b191d21-d1ab-4df8-9e29-6747e121a047"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.406885 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czgp2\" (UniqueName: \"kubernetes.io/projected/7b191d21-d1ab-4df8-9e29-6747e121a047-kube-api-access-czgp2\") pod \"7b191d21-d1ab-4df8-9e29-6747e121a047\" (UID: \"7b191d21-d1ab-4df8-9e29-6747e121a047\") " Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.407339 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b191d21-d1ab-4df8-9e29-6747e121a047-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.411897 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b191d21-d1ab-4df8-9e29-6747e121a047-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7b191d21-d1ab-4df8-9e29-6747e121a047" (UID: "7b191d21-d1ab-4df8-9e29-6747e121a047"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.425243 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b191d21-d1ab-4df8-9e29-6747e121a047-kube-api-access-czgp2" (OuterVolumeSpecName: "kube-api-access-czgp2") pod "7b191d21-d1ab-4df8-9e29-6747e121a047" (UID: "7b191d21-d1ab-4df8-9e29-6747e121a047"). InnerVolumeSpecName "kube-api-access-czgp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.509000 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b191d21-d1ab-4df8-9e29-6747e121a047-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.509039 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czgp2\" (UniqueName: \"kubernetes.io/projected/7b191d21-d1ab-4df8-9e29-6747e121a047-kube-api-access-czgp2\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.574347 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.574746 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn" event={"ID":"7b191d21-d1ab-4df8-9e29-6747e121a047","Type":"ContainerDied","Data":"7951ee9fcd7c0425a1c79749e4a61e7ad1518e62aa1fe63854d7a6c362d59d80"} Dec 10 19:15:18 crc kubenswrapper[4894]: I1210 19:15:18.574798 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7951ee9fcd7c0425a1c79749e4a61e7ad1518e62aa1fe63854d7a6c362d59d80" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.276009 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.576544 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-gl9nv"] Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.577313 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.577364 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:15:19 crc kubenswrapper[4894]: E1210 19:15:19.577424 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b191d21-d1ab-4df8-9e29-6747e121a047" containerName="collect-profiles" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.577467 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b191d21-d1ab-4df8-9e29-6747e121a047" containerName="collect-profiles" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.577780 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b191d21-d1ab-4df8-9e29-6747e121a047" containerName="collect-profiles" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.578752 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.580155 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74b5960f63532f14bffacaae454d958ec29335c9f8adf904077f837019704bae"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.580246 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://74b5960f63532f14bffacaae454d958ec29335c9f8adf904077f837019704bae" gracePeriod=600 Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.581379 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.597821 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gl9nv"] Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.598106 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.629243 4894 generic.go:334] "Generic (PLEG): container finished" podID="88520de0-eedd-40d7-ad33-ee752e8f2ff5" containerID="cf3ec44702652906c442f0b5ba02fc68571f55ed12e5c341d2b7a1a812232f57" exitCode=0 Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.629333 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88520de0-eedd-40d7-ad33-ee752e8f2ff5","Type":"ContainerDied","Data":"cf3ec44702652906c442f0b5ba02fc68571f55ed12e5c341d2b7a1a812232f57"} Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.644327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-ovs-rundir\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.644425 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-ovn-rundir\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.644477 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-combined-ca-bundle\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.644514 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpbnf\" (UniqueName: \"kubernetes.io/projected/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-kube-api-access-mpbnf\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.644573 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.644625 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-config\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.646815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b","Type":"ContainerStarted","Data":"7286128eafcbded3f1d2c1ae9b7ec77db3d29265efc3ef031002c97559c27321"} Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.653070 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f7ec72a7-3043-4044-adfc-7ca2bf0c1749","Type":"ContainerStarted","Data":"6fddef9a9e5629d4a0013bd0c1076e1280da9f41a2b36c8bbe21669eb5ed6697"} Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.712642 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.309402195 podStartE2EDuration="44.712623646s" podCreationTimestamp="2025-12-10 19:14:35 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.108857652 +0000 UTC m=+1173.903705444" lastFinishedPulling="2025-12-10 19:15:18.512079103 +0000 UTC m=+1201.306926895" observedRunningTime="2025-12-10 19:15:19.697143485 +0000 UTC m=+1202.491991277" watchObservedRunningTime="2025-12-10 19:15:19.712623646 +0000 UTC m=+1202.507471448" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.759029 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-ovs-rundir\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.759174 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-ovn-rundir\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.759216 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-combined-ca-bundle\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.759262 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpbnf\" (UniqueName: \"kubernetes.io/projected/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-kube-api-access-mpbnf\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.759328 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.759400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-config\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.760111 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-config\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.760362 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-ovs-rundir\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.762053 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-ovn-rundir\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.786493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.790365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-combined-ca-bundle\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.800561 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpbnf\" (UniqueName: \"kubernetes.io/projected/023ba063-9ce8-45f2-aa5e-9eb7ee33e79c-kube-api-access-mpbnf\") pod \"ovn-controller-metrics-gl9nv\" (UID: \"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c\") " pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.839567 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=15.082182304 podStartE2EDuration="42.839524459s" podCreationTimestamp="2025-12-10 19:14:37 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.765650925 +0000 UTC m=+1173.560498717" lastFinishedPulling="2025-12-10 19:15:18.52299308 +0000 UTC m=+1201.317840872" observedRunningTime="2025-12-10 19:15:19.767130129 +0000 UTC m=+1202.561977921" watchObservedRunningTime="2025-12-10 19:15:19.839524459 +0000 UTC m=+1202.634372251" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.871071 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jnlkn"] Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.871296 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" containerName="dnsmasq-dns" containerID="cri-o://dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a" gracePeriod=10 Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.909912 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gl9nv" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.914742 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-t7sk7"] Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.916678 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.929773 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.962536 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-t7sk7"] Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.987070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.987136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl5cl\" (UniqueName: \"kubernetes.io/projected/389cc9dd-bab9-4ffb-ada8-0c99a2038031-kube-api-access-bl5cl\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.987159 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:19 crc kubenswrapper[4894]: I1210 19:15:19.987233 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-config\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.056523 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tstn2"] Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.056725 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" containerName="dnsmasq-dns" containerID="cri-o://3ff514733815170b51661556993c6e3340f6b6fa15fc1716859f14ea177cf54b" gracePeriod=10 Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.058296 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.089073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.089362 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl5cl\" (UniqueName: \"kubernetes.io/projected/389cc9dd-bab9-4ffb-ada8-0c99a2038031-kube-api-access-bl5cl\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.089383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.089473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-config\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.090332 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-config\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.093379 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.097217 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.127016 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl5cl\" (UniqueName: \"kubernetes.io/projected/389cc9dd-bab9-4ffb-ada8-0c99a2038031-kube-api-access-bl5cl\") pod \"dnsmasq-dns-5bf47b49b7-t7sk7\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.128494 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-dclvk"] Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.132462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.146354 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.156578 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dclvk"] Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.200905 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrp6p\" (UniqueName: \"kubernetes.io/projected/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-kube-api-access-mrp6p\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.200992 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-config\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.201012 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.201070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.201089 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.236470 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.306075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-config\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.306117 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.306219 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.306248 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.306332 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrp6p\" (UniqueName: \"kubernetes.io/projected/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-kube-api-access-mrp6p\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.306940 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-config\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.307165 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.307466 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.307632 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.327316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrp6p\" (UniqueName: \"kubernetes.io/projected/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-kube-api-access-mrp6p\") pod \"dnsmasq-dns-8554648995-dclvk\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.474622 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.566585 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gl9nv"] Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.569726 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.661812 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gl9nv" event={"ID":"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c","Type":"ContainerStarted","Data":"dc6fd5825a6fb2b31a255def81545e7c1f549d719be1a4db86a0182b5fcfab64"} Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.664237 4894 generic.go:334] "Generic (PLEG): container finished" podID="e6752b66-a898-4b95-a065-01c0da2024ef" containerID="dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a" exitCode=0 Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.664272 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" event={"ID":"e6752b66-a898-4b95-a065-01c0da2024ef","Type":"ContainerDied","Data":"dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a"} Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.664288 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" event={"ID":"e6752b66-a898-4b95-a065-01c0da2024ef","Type":"ContainerDied","Data":"1e680624e91378c63bf9ab7a638162748640bd4a41b939bb08772dd42341f4e8"} Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.664308 4894 scope.go:117] "RemoveContainer" containerID="dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.664396 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-jnlkn" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.667492 4894 generic.go:334] "Generic (PLEG): container finished" podID="c41065a9-9e5a-4157-b094-6524bc953df1" containerID="3ff514733815170b51661556993c6e3340f6b6fa15fc1716859f14ea177cf54b" exitCode=0 Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.667558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" event={"ID":"c41065a9-9e5a-4157-b094-6524bc953df1","Type":"ContainerDied","Data":"3ff514733815170b51661556993c6e3340f6b6fa15fc1716859f14ea177cf54b"} Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.669861 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="74b5960f63532f14bffacaae454d958ec29335c9f8adf904077f837019704bae" exitCode=0 Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.670196 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"74b5960f63532f14bffacaae454d958ec29335c9f8adf904077f837019704bae"} Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.670238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"5c62314c4a97570eb40f37c7f81fec8ac3dc6ad9954e24945138dbb31ee1ce97"} Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.700553 4894 scope.go:117] "RemoveContainer" containerID="0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.726242 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-dns-svc\") pod \"e6752b66-a898-4b95-a065-01c0da2024ef\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.726637 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfwdb\" (UniqueName: \"kubernetes.io/projected/e6752b66-a898-4b95-a065-01c0da2024ef-kube-api-access-nfwdb\") pod \"e6752b66-a898-4b95-a065-01c0da2024ef\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.726740 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-config\") pod \"e6752b66-a898-4b95-a065-01c0da2024ef\" (UID: \"e6752b66-a898-4b95-a065-01c0da2024ef\") " Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.732524 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6752b66-a898-4b95-a065-01c0da2024ef-kube-api-access-nfwdb" (OuterVolumeSpecName: "kube-api-access-nfwdb") pod "e6752b66-a898-4b95-a065-01c0da2024ef" (UID: "e6752b66-a898-4b95-a065-01c0da2024ef"). InnerVolumeSpecName "kube-api-access-nfwdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.733642 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfwdb\" (UniqueName: \"kubernetes.io/projected/e6752b66-a898-4b95-a065-01c0da2024ef-kube-api-access-nfwdb\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.738085 4894 scope.go:117] "RemoveContainer" containerID="dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a" Dec 10 19:15:20 crc kubenswrapper[4894]: E1210 19:15:20.745130 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a\": container with ID starting with dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a not found: ID does not exist" containerID="dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.745195 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a"} err="failed to get container status \"dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a\": rpc error: code = NotFound desc = could not find container \"dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a\": container with ID starting with dd3bd09cbfc67b645c88962d17d1d00768309c85357efd7eb2874b7682650d8a not found: ID does not exist" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.745237 4894 scope.go:117] "RemoveContainer" containerID="0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344" Dec 10 19:15:20 crc kubenswrapper[4894]: E1210 19:15:20.747466 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344\": container with ID starting with 0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344 not found: ID does not exist" containerID="0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.747507 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344"} err="failed to get container status \"0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344\": rpc error: code = NotFound desc = could not find container \"0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344\": container with ID starting with 0f63337f3d16f25735b703d0a5b0722689b3e5ecbdf628031d39aa15d4b54344 not found: ID does not exist" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.747532 4894 scope.go:117] "RemoveContainer" containerID="a0a333f57d00f6ace0bfa8b35a6220c82d5630155e1eac3a22921478ab8cbce3" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.788685 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6752b66-a898-4b95-a065-01c0da2024ef" (UID: "e6752b66-a898-4b95-a065-01c0da2024ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.794746 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-config" (OuterVolumeSpecName: "config") pod "e6752b66-a898-4b95-a065-01c0da2024ef" (UID: "e6752b66-a898-4b95-a065-01c0da2024ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.836546 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.836603 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6752b66-a898-4b95-a065-01c0da2024ef-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.859641 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-t7sk7"] Dec 10 19:15:20 crc kubenswrapper[4894]: I1210 19:15:20.966615 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.011674 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dclvk"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.037206 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jnlkn"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.076970 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-jnlkn"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.078194 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.114772 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-t7sk7"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.124375 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.149556 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5dlsv"] Dec 10 19:15:21 crc kubenswrapper[4894]: E1210 19:15:21.149952 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" containerName="init" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.149966 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" containerName="init" Dec 10 19:15:21 crc kubenswrapper[4894]: E1210 19:15:21.149985 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" containerName="dnsmasq-dns" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.149992 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" containerName="dnsmasq-dns" Dec 10 19:15:21 crc kubenswrapper[4894]: E1210 19:15:21.150007 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" containerName="init" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.150015 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" containerName="init" Dec 10 19:15:21 crc kubenswrapper[4894]: E1210 19:15:21.150034 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" containerName="dnsmasq-dns" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.150041 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" containerName="dnsmasq-dns" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.150230 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" containerName="dnsmasq-dns" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.150269 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" containerName="dnsmasq-dns" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.151268 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.219594 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5dlsv"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.251506 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-dns-svc\") pod \"c41065a9-9e5a-4157-b094-6524bc953df1\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.251671 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gnwn\" (UniqueName: \"kubernetes.io/projected/c41065a9-9e5a-4157-b094-6524bc953df1-kube-api-access-6gnwn\") pod \"c41065a9-9e5a-4157-b094-6524bc953df1\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.251795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-config\") pod \"c41065a9-9e5a-4157-b094-6524bc953df1\" (UID: \"c41065a9-9e5a-4157-b094-6524bc953df1\") " Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.252082 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-config\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.252103 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.252163 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.252191 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.252226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn664\" (UniqueName: \"kubernetes.io/projected/115d719b-0a27-44a4-b117-a20211e65090-kube-api-access-sn664\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.264090 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41065a9-9e5a-4157-b094-6524bc953df1-kube-api-access-6gnwn" (OuterVolumeSpecName: "kube-api-access-6gnwn") pod "c41065a9-9e5a-4157-b094-6524bc953df1" (UID: "c41065a9-9e5a-4157-b094-6524bc953df1"). InnerVolumeSpecName "kube-api-access-6gnwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.320980 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-config" (OuterVolumeSpecName: "config") pod "c41065a9-9e5a-4157-b094-6524bc953df1" (UID: "c41065a9-9e5a-4157-b094-6524bc953df1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.348549 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c41065a9-9e5a-4157-b094-6524bc953df1" (UID: "c41065a9-9e5a-4157-b094-6524bc953df1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.353981 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354040 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn664\" (UniqueName: \"kubernetes.io/projected/115d719b-0a27-44a4-b117-a20211e65090-kube-api-access-sn664\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354152 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-config\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354173 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354277 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354288 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c41065a9-9e5a-4157-b094-6524bc953df1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.354297 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gnwn\" (UniqueName: \"kubernetes.io/projected/c41065a9-9e5a-4157-b094-6524bc953df1-kube-api-access-6gnwn\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.355052 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.355050 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.355224 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.355550 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-config\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.374123 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn664\" (UniqueName: \"kubernetes.io/projected/115d719b-0a27-44a4-b117-a20211e65090-kube-api-access-sn664\") pod \"dnsmasq-dns-b8fbc5445-5dlsv\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.478576 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.558222 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6752b66-a898-4b95-a065-01c0da2024ef" path="/var/lib/kubelet/pods/e6752b66-a898-4b95-a065-01c0da2024ef/volumes" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.695646 4894 generic.go:334] "Generic (PLEG): container finished" podID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerID="ef94644c9662f85171d2ed5a61b5c3d44442e22aefcdaa88814d7c4529769bd2" exitCode=0 Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.695754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dclvk" event={"ID":"b09cbd3e-9108-4d01-a4f0-abf9e680b16d","Type":"ContainerDied","Data":"ef94644c9662f85171d2ed5a61b5c3d44442e22aefcdaa88814d7c4529769bd2"} Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.696135 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dclvk" event={"ID":"b09cbd3e-9108-4d01-a4f0-abf9e680b16d","Type":"ContainerStarted","Data":"0af1746bfefca88e65dca4d555c3a4f0b03436fdf7b446e7346ca7ba04cdd252"} Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.697967 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gl9nv" event={"ID":"023ba063-9ce8-45f2-aa5e-9eb7ee33e79c","Type":"ContainerStarted","Data":"c35db67104e0aa276f94a0359962e2f74ca78e3574a2130258f36ababc83a9fd"} Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.706628 4894 generic.go:334] "Generic (PLEG): container finished" podID="389cc9dd-bab9-4ffb-ada8-0c99a2038031" containerID="11c3422b6ef0f1eb7b3a4a0409d6f1aadbfa47fdc5a447e4a257429d22d1c750" exitCode=0 Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.706731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" event={"ID":"389cc9dd-bab9-4ffb-ada8-0c99a2038031","Type":"ContainerDied","Data":"11c3422b6ef0f1eb7b3a4a0409d6f1aadbfa47fdc5a447e4a257429d22d1c750"} Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.706767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" event={"ID":"389cc9dd-bab9-4ffb-ada8-0c99a2038031","Type":"ContainerStarted","Data":"11987f2152494efaac98d9f88a91d9bc2396e9ed5d66ffba28f79883fd066a0f"} Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.711504 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" event={"ID":"c41065a9-9e5a-4157-b094-6524bc953df1","Type":"ContainerDied","Data":"9abe5a81331b6ba03b3de64d88f4c839cb230acc83bc099df35c4067b5dbb870"} Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.711527 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tstn2" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.711567 4894 scope.go:117] "RemoveContainer" containerID="3ff514733815170b51661556993c6e3340f6b6fa15fc1716859f14ea177cf54b" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.711985 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.780724 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.781655 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-gl9nv" podStartSLOduration=2.781639847 podStartE2EDuration="2.781639847s" podCreationTimestamp="2025-12-10 19:15:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:21.750320425 +0000 UTC m=+1204.545168217" watchObservedRunningTime="2025-12-10 19:15:21.781639847 +0000 UTC m=+1204.576487639" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.799463 4894 scope.go:117] "RemoveContainer" containerID="effbb40c043be21dc23390924b40b151fa8e72f242b74fa45a456da83010b865" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.800096 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tstn2"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.801733 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.801996 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.813117 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tstn2"] Dec 10 19:15:21 crc kubenswrapper[4894]: I1210 19:15:21.991898 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5dlsv"] Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.058489 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.158561 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.260403 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.261124 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="389cc9dd-bab9-4ffb-ada8-0c99a2038031" containerName="init" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.261137 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="389cc9dd-bab9-4ffb-ada8-0c99a2038031" containerName="init" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.261326 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="389cc9dd-bab9-4ffb-ada8-0c99a2038031" containerName="init" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.270465 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.273030 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.273174 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl5cl\" (UniqueName: \"kubernetes.io/projected/389cc9dd-bab9-4ffb-ada8-0c99a2038031-kube-api-access-bl5cl\") pod \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.273233 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-dns-svc\") pod \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.273428 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-config\") pod \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.273477 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-ovsdbserver-nb\") pod \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\" (UID: \"389cc9dd-bab9-4ffb-ada8-0c99a2038031\") " Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.275585 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-rm2gd" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.275627 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.276157 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.282966 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/389cc9dd-bab9-4ffb-ada8-0c99a2038031-kube-api-access-bl5cl" (OuterVolumeSpecName: "kube-api-access-bl5cl") pod "389cc9dd-bab9-4ffb-ada8-0c99a2038031" (UID: "389cc9dd-bab9-4ffb-ada8-0c99a2038031"). InnerVolumeSpecName "kube-api-access-bl5cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.301411 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "389cc9dd-bab9-4ffb-ada8-0c99a2038031" (UID: "389cc9dd-bab9-4ffb-ada8-0c99a2038031"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.305628 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "389cc9dd-bab9-4ffb-ada8-0c99a2038031" (UID: "389cc9dd-bab9-4ffb-ada8-0c99a2038031"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.313309 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.334107 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-config" (OuterVolumeSpecName: "config") pod "389cc9dd-bab9-4ffb-ada8-0c99a2038031" (UID: "389cc9dd-bab9-4ffb-ada8-0c99a2038031"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375031 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375088 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375116 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-lock\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375165 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkn2k\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-kube-api-access-qkn2k\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375206 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-cache\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375266 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375277 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375286 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl5cl\" (UniqueName: \"kubernetes.io/projected/389cc9dd-bab9-4ffb-ada8-0c99a2038031-kube-api-access-bl5cl\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.375295 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389cc9dd-bab9-4ffb-ada8-0c99a2038031-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.472007 4894 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 10 19:15:22 crc kubenswrapper[4894]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/b09cbd3e-9108-4d01-a4f0-abf9e680b16d/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 10 19:15:22 crc kubenswrapper[4894]: > podSandboxID="0af1746bfefca88e65dca4d555c3a4f0b03436fdf7b446e7346ca7ba04cdd252" Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.472248 4894 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 10 19:15:22 crc kubenswrapper[4894]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n654h99h64ch5dbh6dh555h587h64bh5cfh647h5fdh57ch679h9h597h5f5hbch59bh54fh575h566h667h586h5f5h65ch5bch57h68h65ch58bh694h5cfq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mrp6p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8554648995-dclvk_openstack(b09cbd3e-9108-4d01-a4f0-abf9e680b16d): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/b09cbd3e-9108-4d01-a4f0-abf9e680b16d/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 10 19:15:22 crc kubenswrapper[4894]: > logger="UnhandledError" Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.473461 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/b09cbd3e-9108-4d01-a4f0-abf9e680b16d/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-8554648995-dclvk" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.476740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.476803 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.476833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-lock\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.476921 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkn2k\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-kube-api-access-qkn2k\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.476981 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.477002 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:15:22 crc kubenswrapper[4894]: E1210 19:15:22.477057 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift podName:5eac36f4-3cbc-48c1-9c6a-4e79de593d04 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:22.977038167 +0000 UTC m=+1205.771885959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift") pod "swift-storage-0" (UID: "5eac36f4-3cbc-48c1-9c6a-4e79de593d04") : configmap "swift-ring-files" not found Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.476981 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-cache\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.477634 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-lock\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.477862 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-cache\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.482222 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.482264 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/390fdaf865d4c8a159f68a19d51ed4682e65c30595fb04eb182b21f094776e22/globalmount\"" pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.493773 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkn2k\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-kube-api-access-qkn2k\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.529974 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ca25e1e-de25-4ed5-a7c6-5b38da656ce1\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.725563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" event={"ID":"389cc9dd-bab9-4ffb-ada8-0c99a2038031","Type":"ContainerDied","Data":"11987f2152494efaac98d9f88a91d9bc2396e9ed5d66ffba28f79883fd066a0f"} Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.725613 4894 scope.go:117] "RemoveContainer" containerID="11c3422b6ef0f1eb7b3a4a0409d6f1aadbfa47fdc5a447e4a257429d22d1c750" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.725730 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-t7sk7" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.731161 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerID="08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f" exitCode=0 Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.731236 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerDied","Data":"08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f"} Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.738870 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" event={"ID":"115d719b-0a27-44a4-b117-a20211e65090","Type":"ContainerStarted","Data":"2fc020f0584646bb06af9f8ba60a0641f7c1377bce06d1fddb18e71cde3feacd"} Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.807679 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.862541 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-t7sk7"] Dec 10 19:15:22 crc kubenswrapper[4894]: I1210 19:15:22.880625 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-t7sk7"] Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.008476 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:23 crc kubenswrapper[4894]: E1210 19:15:23.008971 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:15:23 crc kubenswrapper[4894]: E1210 19:15:23.008985 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:15:23 crc kubenswrapper[4894]: E1210 19:15:23.009026 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift podName:5eac36f4-3cbc-48c1-9c6a-4e79de593d04 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:24.00901188 +0000 UTC m=+1206.803859672 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift") pod "swift-storage-0" (UID: "5eac36f4-3cbc-48c1-9c6a-4e79de593d04") : configmap "swift-ring-files" not found Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.166836 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.168732 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.171837 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.172156 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.172297 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8clsp" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.172500 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.191677 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222430 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222536 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222590 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222691 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr8bk\" (UniqueName: \"kubernetes.io/projected/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-kube-api-access-lr8bk\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222729 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-scripts\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-config\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.222890 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.324736 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr8bk\" (UniqueName: \"kubernetes.io/projected/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-kube-api-access-lr8bk\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.324784 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-scripts\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.324810 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-config\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.324891 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.324931 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.324975 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.325010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.327311 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-config\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.328485 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.328838 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-scripts\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.331086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.334021 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.334435 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.344239 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr8bk\" (UniqueName: \"kubernetes.io/projected/ec34ebe8-9b05-4fc8-9866-d6cb02e61504-kube-api-access-lr8bk\") pod \"ovn-northd-0\" (UID: \"ec34ebe8-9b05-4fc8-9866-d6cb02e61504\") " pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.532812 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.553651 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="389cc9dd-bab9-4ffb-ada8-0c99a2038031" path="/var/lib/kubelet/pods/389cc9dd-bab9-4ffb-ada8-0c99a2038031/volumes" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.554435 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c41065a9-9e5a-4157-b094-6524bc953df1" path="/var/lib/kubelet/pods/c41065a9-9e5a-4157-b094-6524bc953df1/volumes" Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.763822 4894 generic.go:334] "Generic (PLEG): container finished" podID="115d719b-0a27-44a4-b117-a20211e65090" containerID="e680f26706f00ac782aefd533381d1fbdff67881cf2e4d3812e5e07eacaf1615" exitCode=0 Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.763911 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" event={"ID":"115d719b-0a27-44a4-b117-a20211e65090","Type":"ContainerDied","Data":"e680f26706f00ac782aefd533381d1fbdff67881cf2e4d3812e5e07eacaf1615"} Dec 10 19:15:23 crc kubenswrapper[4894]: I1210 19:15:23.769549 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0297c834-eee7-4414-ba18-b4ee6b95561f","Type":"ContainerStarted","Data":"db8c55a2f87977f156adeaf9ff9579e44e99acde44030ccad21bbb204517ec9b"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.043188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:24 crc kubenswrapper[4894]: E1210 19:15:24.043371 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:15:24 crc kubenswrapper[4894]: E1210 19:15:24.043744 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:15:24 crc kubenswrapper[4894]: E1210 19:15:24.043797 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift podName:5eac36f4-3cbc-48c1-9c6a-4e79de593d04 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:26.043780342 +0000 UTC m=+1208.838628134 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift") pod "swift-storage-0" (UID: "5eac36f4-3cbc-48c1-9c6a-4e79de593d04") : configmap "swift-ring-files" not found Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.097240 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.779615 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88520de0-eedd-40d7-ad33-ee752e8f2ff5","Type":"ContainerStarted","Data":"331c6555ccc297ab1426a732e42249a9a8478c8c2ed82fa5d07aa11a81010bfe"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.783618 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" event={"ID":"115d719b-0a27-44a4-b117-a20211e65090","Type":"ContainerStarted","Data":"3eac62ee0066a6901bf1400403d95b9e486a4c214e4f28d0b09fe699eacb2646"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.783754 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.784950 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" event={"ID":"bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae","Type":"ContainerStarted","Data":"bccc817e2ef90393fa20041bea1630f4439a4f418181acc8c04a5868ec56cfae"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.785279 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.787552 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dclvk" event={"ID":"b09cbd3e-9108-4d01-a4f0-abf9e680b16d","Type":"ContainerStarted","Data":"dd136135e1ba057697dc1949a87d92ef89b663129c48a455f95efc8878a3aa4a"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.787713 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.788782 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ec34ebe8-9b05-4fc8-9866-d6cb02e61504","Type":"ContainerStarted","Data":"8e50d1f5cf34d7edcd832958bac382b9282c91be907a1af4d39c33387cdf8984"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.805175 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" podStartSLOduration=3.805160737 podStartE2EDuration="3.805160737s" podCreationTimestamp="2025-12-10 19:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:24.799144693 +0000 UTC m=+1207.593992495" watchObservedRunningTime="2025-12-10 19:15:24.805160737 +0000 UTC m=+1207.600008519" Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.808114 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab75b963-c89e-4871-8cb1-180dbaaf2714","Type":"ContainerStarted","Data":"74cc0a8e43d22c0af8f41c9b23afd66bc35d52380bd0a137af51f727db786ea0"} Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.824496 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" podStartSLOduration=-9223371995.030306 podStartE2EDuration="41.824469722s" podCreationTimestamp="2025-12-10 19:14:43 +0000 UTC" firstStartedPulling="2025-12-10 19:14:51.044683966 +0000 UTC m=+1173.839531758" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:24.816056663 +0000 UTC m=+1207.610904475" watchObservedRunningTime="2025-12-10 19:15:24.824469722 +0000 UTC m=+1207.619317514" Dec 10 19:15:24 crc kubenswrapper[4894]: I1210 19:15:24.830270 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-dclvk" podStartSLOduration=4.830250469 podStartE2EDuration="4.830250469s" podCreationTimestamp="2025-12-10 19:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:24.829247022 +0000 UTC m=+1207.624094814" watchObservedRunningTime="2025-12-10 19:15:24.830250469 +0000 UTC m=+1207.625098261" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.084803 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:26 crc kubenswrapper[4894]: E1210 19:15:26.085023 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:15:26 crc kubenswrapper[4894]: E1210 19:15:26.085264 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:15:26 crc kubenswrapper[4894]: E1210 19:15:26.085340 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift podName:5eac36f4-3cbc-48c1-9c6a-4e79de593d04 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:30.085321406 +0000 UTC m=+1212.880169198 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift") pod "swift-storage-0" (UID: "5eac36f4-3cbc-48c1-9c6a-4e79de593d04") : configmap "swift-ring-files" not found Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.208820 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-mxbcj"] Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.210124 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.212015 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.212467 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.213364 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.224359 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mxbcj"] Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392435 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-dispersionconf\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392487 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-combined-ca-bundle\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392523 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scpwk\" (UniqueName: \"kubernetes.io/projected/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-kube-api-access-scpwk\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392578 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-etc-swift\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392624 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-ring-data-devices\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392697 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-scripts\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.392741 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-swiftconf\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-ring-data-devices\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495162 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-scripts\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495221 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-swiftconf\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495338 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-dispersionconf\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495371 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-combined-ca-bundle\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495399 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scpwk\" (UniqueName: \"kubernetes.io/projected/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-kube-api-access-scpwk\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-etc-swift\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495794 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-ring-data-devices\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495866 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-etc-swift\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.495889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-scripts\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.690491 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-dispersionconf\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.691401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-combined-ca-bundle\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.692376 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-swiftconf\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.700520 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scpwk\" (UniqueName: \"kubernetes.io/projected/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-kube-api-access-scpwk\") pod \"swift-ring-rebalance-mxbcj\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:26 crc kubenswrapper[4894]: I1210 19:15:26.830457 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.391602 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mxbcj"] Dec 10 19:15:27 crc kubenswrapper[4894]: W1210 19:15:27.396920 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfee50bb9_aa22_44b0_b596_c3f64d22c8a3.slice/crio-b9a6bdda234af1c38289fc37d4b914ddb10ec37ad97bd7c945d6a88134aa4a3d WatchSource:0}: Error finding container b9a6bdda234af1c38289fc37d4b914ddb10ec37ad97bd7c945d6a88134aa4a3d: Status 404 returned error can't find the container with id b9a6bdda234af1c38289fc37d4b914ddb10ec37ad97bd7c945d6a88134aa4a3d Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.842582 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"88520de0-eedd-40d7-ad33-ee752e8f2ff5","Type":"ContainerStarted","Data":"d06940ad0ea334659020697387420c750d5eed3d50a1fd4b7d982c3f5de8d2cd"} Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.844114 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.845938 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.846392 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ec34ebe8-9b05-4fc8-9866-d6cb02e61504","Type":"ContainerStarted","Data":"44877e65d63fc3ce4faf3e2163c0c63ab707035aeed4b645d7c461d78f6f837b"} Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.846416 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ec34ebe8-9b05-4fc8-9866-d6cb02e61504","Type":"ContainerStarted","Data":"7e9ae119bce178226a878a33662eb696930e382cd64ba7430d39ee78a52c5e21"} Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.846451 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.848229 4894 generic.go:334] "Generic (PLEG): container finished" podID="0297c834-eee7-4414-ba18-b4ee6b95561f" containerID="db8c55a2f87977f156adeaf9ff9579e44e99acde44030ccad21bbb204517ec9b" exitCode=0 Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.848280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0297c834-eee7-4414-ba18-b4ee6b95561f","Type":"ContainerDied","Data":"db8c55a2f87977f156adeaf9ff9579e44e99acde44030ccad21bbb204517ec9b"} Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.849968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mxbcj" event={"ID":"fee50bb9-aa22-44b0-b596-c3f64d22c8a3","Type":"ContainerStarted","Data":"b9a6bdda234af1c38289fc37d4b914ddb10ec37ad97bd7c945d6a88134aa4a3d"} Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.867046 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=23.083914413 podStartE2EDuration="56.86702242s" podCreationTimestamp="2025-12-10 19:14:31 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.028203821 +0000 UTC m=+1172.823051613" lastFinishedPulling="2025-12-10 19:15:23.811311828 +0000 UTC m=+1206.606159620" observedRunningTime="2025-12-10 19:15:27.86153488 +0000 UTC m=+1210.656382672" watchObservedRunningTime="2025-12-10 19:15:27.86702242 +0000 UTC m=+1210.661870202" Dec 10 19:15:27 crc kubenswrapper[4894]: I1210 19:15:27.888776 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.108278523 podStartE2EDuration="4.888758681s" podCreationTimestamp="2025-12-10 19:15:23 +0000 UTC" firstStartedPulling="2025-12-10 19:15:24.116767948 +0000 UTC m=+1206.911615740" lastFinishedPulling="2025-12-10 19:15:26.897248106 +0000 UTC m=+1209.692095898" observedRunningTime="2025-12-10 19:15:27.885198034 +0000 UTC m=+1210.680045816" watchObservedRunningTime="2025-12-10 19:15:27.888758681 +0000 UTC m=+1210.683606493" Dec 10 19:15:28 crc kubenswrapper[4894]: I1210 19:15:28.859446 4894 generic.go:334] "Generic (PLEG): container finished" podID="ab75b963-c89e-4871-8cb1-180dbaaf2714" containerID="74cc0a8e43d22c0af8f41c9b23afd66bc35d52380bd0a137af51f727db786ea0" exitCode=0 Dec 10 19:15:28 crc kubenswrapper[4894]: I1210 19:15:28.859557 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab75b963-c89e-4871-8cb1-180dbaaf2714","Type":"ContainerDied","Data":"74cc0a8e43d22c0af8f41c9b23afd66bc35d52380bd0a137af51f727db786ea0"} Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.164998 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:30 crc kubenswrapper[4894]: E1210 19:15:30.165321 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:15:30 crc kubenswrapper[4894]: E1210 19:15:30.165394 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:15:30 crc kubenswrapper[4894]: E1210 19:15:30.165522 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift podName:5eac36f4-3cbc-48c1-9c6a-4e79de593d04 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:38.165494334 +0000 UTC m=+1220.960342146 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift") pod "swift-storage-0" (UID: "5eac36f4-3cbc-48c1-9c6a-4e79de593d04") : configmap "swift-ring-files" not found Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.477108 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.890534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerStarted","Data":"ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec"} Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.892496 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d","Type":"ContainerStarted","Data":"9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2"} Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.892665 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.895465 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab75b963-c89e-4871-8cb1-180dbaaf2714","Type":"ContainerStarted","Data":"e5a27f3c144ca94e2e60f053167fb68aa03f7d1bd92fb2d0fb3830ad11ab10d8"} Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.898086 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0297c834-eee7-4414-ba18-b4ee6b95561f","Type":"ContainerStarted","Data":"54b4073ad84adbd130965d8f98a8983d29aabb02c9c5b9cb8b814ef47cf275dd"} Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.911324 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=21.400523298 podStartE2EDuration="1m0.911309335s" podCreationTimestamp="2025-12-10 19:14:30 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.598397055 +0000 UTC m=+1173.393244847" lastFinishedPulling="2025-12-10 19:15:30.109183082 +0000 UTC m=+1212.904030884" observedRunningTime="2025-12-10 19:15:30.907662276 +0000 UTC m=+1213.702510068" watchObservedRunningTime="2025-12-10 19:15:30.911309335 +0000 UTC m=+1213.706157137" Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.932385 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=32.755200094 podStartE2EDuration="1m4.932366988s" podCreationTimestamp="2025-12-10 19:14:26 +0000 UTC" firstStartedPulling="2025-12-10 19:14:49.976978318 +0000 UTC m=+1172.771826110" lastFinishedPulling="2025-12-10 19:15:22.154145212 +0000 UTC m=+1204.948993004" observedRunningTime="2025-12-10 19:15:30.926455816 +0000 UTC m=+1213.721303618" watchObservedRunningTime="2025-12-10 19:15:30.932366988 +0000 UTC m=+1213.727214780" Dec 10 19:15:30 crc kubenswrapper[4894]: I1210 19:15:30.950651 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371972.904142 podStartE2EDuration="1m3.950634085s" podCreationTimestamp="2025-12-10 19:14:27 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.668199613 +0000 UTC m=+1173.463047405" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:30.944765724 +0000 UTC m=+1213.739613527" watchObservedRunningTime="2025-12-10 19:15:30.950634085 +0000 UTC m=+1213.745481877" Dec 10 19:15:31 crc kubenswrapper[4894]: E1210 19:15:31.230455 4894 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.201:42866->38.102.83.201:41037: write tcp 38.102.83.201:42866->38.102.83.201:41037: write: broken pipe Dec 10 19:15:31 crc kubenswrapper[4894]: I1210 19:15:31.479986 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:15:31 crc kubenswrapper[4894]: I1210 19:15:31.558885 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dclvk"] Dec 10 19:15:31 crc kubenswrapper[4894]: I1210 19:15:31.559461 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-dclvk" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="dnsmasq-dns" containerID="cri-o://dd136135e1ba057697dc1949a87d92ef89b663129c48a455f95efc8878a3aa4a" gracePeriod=10 Dec 10 19:15:31 crc kubenswrapper[4894]: I1210 19:15:31.906597 4894 generic.go:334] "Generic (PLEG): container finished" podID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerID="dd136135e1ba057697dc1949a87d92ef89b663129c48a455f95efc8878a3aa4a" exitCode=0 Dec 10 19:15:31 crc kubenswrapper[4894]: I1210 19:15:31.906687 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dclvk" event={"ID":"b09cbd3e-9108-4d01-a4f0-abf9e680b16d","Type":"ContainerDied","Data":"dd136135e1ba057697dc1949a87d92ef89b663129c48a455f95efc8878a3aa4a"} Dec 10 19:15:33 crc kubenswrapper[4894]: I1210 19:15:33.927705 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerStarted","Data":"d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5"} Dec 10 19:15:33 crc kubenswrapper[4894]: I1210 19:15:33.949096 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-66dfd9bb-4qx65" Dec 10 19:15:34 crc kubenswrapper[4894]: I1210 19:15:34.257879 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h" Dec 10 19:15:35 crc kubenswrapper[4894]: I1210 19:15:35.188300 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 10 19:15:35 crc kubenswrapper[4894]: I1210 19:15:35.294159 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 10 19:15:35 crc kubenswrapper[4894]: I1210 19:15:35.413535 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="da36895a-2989-4110-88a2-6bf0eb5d2f5d" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.783176 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.906120 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-nb\") pod \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.906169 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-sb\") pod \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.906244 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-config\") pod \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.906288 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrp6p\" (UniqueName: \"kubernetes.io/projected/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-kube-api-access-mrp6p\") pod \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.906409 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc\") pod \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.911249 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-kube-api-access-mrp6p" (OuterVolumeSpecName: "kube-api-access-mrp6p") pod "b09cbd3e-9108-4d01-a4f0-abf9e680b16d" (UID: "b09cbd3e-9108-4d01-a4f0-abf9e680b16d"). InnerVolumeSpecName "kube-api-access-mrp6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.957079 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-dclvk" event={"ID":"b09cbd3e-9108-4d01-a4f0-abf9e680b16d","Type":"ContainerDied","Data":"0af1746bfefca88e65dca4d555c3a4f0b03436fdf7b446e7346ca7ba04cdd252"} Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.957145 4894 scope.go:117] "RemoveContainer" containerID="dd136135e1ba057697dc1949a87d92ef89b663129c48a455f95efc8878a3aa4a" Dec 10 19:15:36 crc kubenswrapper[4894]: I1210 19:15:36.957169 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-dclvk" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.008839 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrp6p\" (UniqueName: \"kubernetes.io/projected/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-kube-api-access-mrp6p\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.012107 4894 scope.go:117] "RemoveContainer" containerID="ef94644c9662f85171d2ed5a61b5c3d44442e22aefcdaa88814d7c4529769bd2" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.016542 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-config" (OuterVolumeSpecName: "config") pod "b09cbd3e-9108-4d01-a4f0-abf9e680b16d" (UID: "b09cbd3e-9108-4d01-a4f0-abf9e680b16d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:37 crc kubenswrapper[4894]: E1210 19:15:37.016991 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc podName:b09cbd3e-9108-4d01-a4f0-abf9e680b16d nodeName:}" failed. No retries permitted until 2025-12-10 19:15:37.516958759 +0000 UTC m=+1220.311806541 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc") pod "b09cbd3e-9108-4d01-a4f0-abf9e680b16d" (UID: "b09cbd3e-9108-4d01-a4f0-abf9e680b16d") : error deleting /var/lib/kubelet/pods/b09cbd3e-9108-4d01-a4f0-abf9e680b16d/volume-subpaths: remove /var/lib/kubelet/pods/b09cbd3e-9108-4d01-a4f0-abf9e680b16d/volume-subpaths: no such file or directory Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.017115 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b09cbd3e-9108-4d01-a4f0-abf9e680b16d" (UID: "b09cbd3e-9108-4d01-a4f0-abf9e680b16d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.017296 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b09cbd3e-9108-4d01-a4f0-abf9e680b16d" (UID: "b09cbd3e-9108-4d01-a4f0-abf9e680b16d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.109988 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.110020 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.110029 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.618188 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc\") pod \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\" (UID: \"b09cbd3e-9108-4d01-a4f0-abf9e680b16d\") " Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.618560 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b09cbd3e-9108-4d01-a4f0-abf9e680b16d" (UID: "b09cbd3e-9108-4d01-a4f0-abf9e680b16d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.618833 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b09cbd3e-9108-4d01-a4f0-abf9e680b16d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.833548 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.834045 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.891183 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dclvk"] Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.899307 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-dclvk"] Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.972514 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mxbcj" event={"ID":"fee50bb9-aa22-44b0-b596-c3f64d22c8a3","Type":"ContainerStarted","Data":"50b727ad86634be3d490dab3919eb1c1087a93a8e05c3e6ae617678aed7d94a0"} Dec 10 19:15:37 crc kubenswrapper[4894]: I1210 19:15:37.998588 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-mxbcj" podStartSLOduration=2.7072921 podStartE2EDuration="11.998573715s" podCreationTimestamp="2025-12-10 19:15:26 +0000 UTC" firstStartedPulling="2025-12-10 19:15:27.39914414 +0000 UTC m=+1210.193991932" lastFinishedPulling="2025-12-10 19:15:36.690425755 +0000 UTC m=+1219.485273547" observedRunningTime="2025-12-10 19:15:37.996775686 +0000 UTC m=+1220.791623498" watchObservedRunningTime="2025-12-10 19:15:37.998573715 +0000 UTC m=+1220.793421507" Dec 10 19:15:38 crc kubenswrapper[4894]: I1210 19:15:38.229997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:38 crc kubenswrapper[4894]: E1210 19:15:38.230172 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:15:38 crc kubenswrapper[4894]: E1210 19:15:38.230189 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:15:38 crc kubenswrapper[4894]: E1210 19:15:38.230233 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift podName:5eac36f4-3cbc-48c1-9c6a-4e79de593d04 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:54.230220608 +0000 UTC m=+1237.025068400 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift") pod "swift-storage-0" (UID: "5eac36f4-3cbc-48c1-9c6a-4e79de593d04") : configmap "swift-ring-files" not found Dec 10 19:15:38 crc kubenswrapper[4894]: I1210 19:15:38.595306 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 10 19:15:39 crc kubenswrapper[4894]: I1210 19:15:39.151102 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 10 19:15:39 crc kubenswrapper[4894]: I1210 19:15:39.151157 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 10 19:15:39 crc kubenswrapper[4894]: I1210 19:15:39.249722 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 10 19:15:39 crc kubenswrapper[4894]: I1210 19:15:39.557526 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" path="/var/lib/kubelet/pods/b09cbd3e-9108-4d01-a4f0-abf9e680b16d/volumes" Dec 10 19:15:40 crc kubenswrapper[4894]: I1210 19:15:40.128449 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 10 19:15:40 crc kubenswrapper[4894]: I1210 19:15:40.476670 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-dclvk" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Dec 10 19:15:40 crc kubenswrapper[4894]: I1210 19:15:40.793875 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 10 19:15:40 crc kubenswrapper[4894]: I1210 19:15:40.871204 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 10 19:15:41 crc kubenswrapper[4894]: I1210 19:15:41.000369 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerStarted","Data":"4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602"} Dec 10 19:15:41 crc kubenswrapper[4894]: I1210 19:15:41.002888 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 19:15:41 crc kubenswrapper[4894]: I1210 19:15:41.029830 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.708407714 podStartE2EDuration="1m11.029815166s" podCreationTimestamp="2025-12-10 19:14:30 +0000 UTC" firstStartedPulling="2025-12-10 19:14:50.633751327 +0000 UTC m=+1173.428599119" lastFinishedPulling="2025-12-10 19:15:39.955158769 +0000 UTC m=+1222.750006571" observedRunningTime="2025-12-10 19:15:41.024463372 +0000 UTC m=+1223.819311194" watchObservedRunningTime="2025-12-10 19:15:41.029815166 +0000 UTC m=+1223.824662958" Dec 10 19:15:42 crc kubenswrapper[4894]: I1210 19:15:42.628878 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:44 crc kubenswrapper[4894]: I1210 19:15:44.252629 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-795fd8f8cc-rhw5x" Dec 10 19:15:45 crc kubenswrapper[4894]: I1210 19:15:45.040023 4894 generic.go:334] "Generic (PLEG): container finished" podID="fee50bb9-aa22-44b0-b596-c3f64d22c8a3" containerID="50b727ad86634be3d490dab3919eb1c1087a93a8e05c3e6ae617678aed7d94a0" exitCode=0 Dec 10 19:15:45 crc kubenswrapper[4894]: I1210 19:15:45.040061 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mxbcj" event={"ID":"fee50bb9-aa22-44b0-b596-c3f64d22c8a3","Type":"ContainerDied","Data":"50b727ad86634be3d490dab3919eb1c1087a93a8e05c3e6ae617678aed7d94a0"} Dec 10 19:15:45 crc kubenswrapper[4894]: I1210 19:15:45.408254 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="da36895a-2989-4110-88a2-6bf0eb5d2f5d" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:15:45 crc kubenswrapper[4894]: I1210 19:15:45.875794 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-h2552" podUID="5fd5a07b-caf3-44d6-a9d4-e50e5be1819c" containerName="ovn-controller" probeResult="failure" output=< Dec 10 19:15:45 crc kubenswrapper[4894]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 19:15:45 crc kubenswrapper[4894]: > Dec 10 19:15:45 crc kubenswrapper[4894]: I1210 19:15:45.943463 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:15:45 crc kubenswrapper[4894]: I1210 19:15:45.950631 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pdwjv" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.176368 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-h2552-config-twfpr"] Dec 10 19:15:46 crc kubenswrapper[4894]: E1210 19:15:46.176708 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="init" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.187462 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="init" Dec 10 19:15:46 crc kubenswrapper[4894]: E1210 19:15:46.187587 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="dnsmasq-dns" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.187661 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="dnsmasq-dns" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.188120 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b09cbd3e-9108-4d01-a4f0-abf9e680b16d" containerName="dnsmasq-dns" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.188736 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h2552-config-twfpr"] Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.188906 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.190734 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.280523 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp7vn\" (UniqueName: \"kubernetes.io/projected/253af012-5d33-45ed-b30c-c2d114635ab3-kube-api-access-wp7vn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.280618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.280651 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-scripts\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.280678 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-additional-scripts\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.280797 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run-ovn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.280815 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-log-ovn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382369 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp7vn\" (UniqueName: \"kubernetes.io/projected/253af012-5d33-45ed-b30c-c2d114635ab3-kube-api-access-wp7vn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382707 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382758 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-scripts\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-additional-scripts\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382827 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run-ovn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382853 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-log-ovn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.382968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.383004 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run-ovn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.383033 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-log-ovn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.383608 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-additional-scripts\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.384599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-scripts\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.402947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp7vn\" (UniqueName: \"kubernetes.io/projected/253af012-5d33-45ed-b30c-c2d114635ab3-kube-api-access-wp7vn\") pod \"ovn-controller-h2552-config-twfpr\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.486871 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.530278 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587360 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-scripts\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-ring-data-devices\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587629 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-dispersionconf\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587696 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-etc-swift\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587719 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scpwk\" (UniqueName: \"kubernetes.io/projected/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-kube-api-access-scpwk\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587772 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-swiftconf\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.587802 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-combined-ca-bundle\") pod \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\" (UID: \"fee50bb9-aa22-44b0-b596-c3f64d22c8a3\") " Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.588392 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.589054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.594153 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-kube-api-access-scpwk" (OuterVolumeSpecName: "kube-api-access-scpwk") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "kube-api-access-scpwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.599570 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.620511 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-scripts" (OuterVolumeSpecName: "scripts") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.627572 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.629026 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fee50bb9-aa22-44b0-b596-c3f64d22c8a3" (UID: "fee50bb9-aa22-44b0-b596-c3f64d22c8a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689786 4894 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689816 4894 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689827 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scpwk\" (UniqueName: \"kubernetes.io/projected/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-kube-api-access-scpwk\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689839 4894 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689860 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689869 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.689876 4894 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fee50bb9-aa22-44b0-b596-c3f64d22c8a3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:46 crc kubenswrapper[4894]: I1210 19:15:46.980986 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h2552-config-twfpr"] Dec 10 19:15:46 crc kubenswrapper[4894]: W1210 19:15:46.986466 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod253af012_5d33_45ed_b30c_c2d114635ab3.slice/crio-ea1af466c8aea196c014326cc3abbee95df49fae3208536ee0e8110f0438e3a3 WatchSource:0}: Error finding container ea1af466c8aea196c014326cc3abbee95df49fae3208536ee0e8110f0438e3a3: Status 404 returned error can't find the container with id ea1af466c8aea196c014326cc3abbee95df49fae3208536ee0e8110f0438e3a3 Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.056827 4894 generic.go:334] "Generic (PLEG): container finished" podID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerID="f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e" exitCode=0 Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.056890 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8f81cda0-aeed-41c1-9941-5f1ba92a88c5","Type":"ContainerDied","Data":"f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e"} Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.061504 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mxbcj" Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.061518 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mxbcj" event={"ID":"fee50bb9-aa22-44b0-b596-c3f64d22c8a3","Type":"ContainerDied","Data":"b9a6bdda234af1c38289fc37d4b914ddb10ec37ad97bd7c945d6a88134aa4a3d"} Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.061552 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9a6bdda234af1c38289fc37d4b914ddb10ec37ad97bd7c945d6a88134aa4a3d" Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.065415 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-twfpr" event={"ID":"253af012-5d33-45ed-b30c-c2d114635ab3","Type":"ContainerStarted","Data":"ea1af466c8aea196c014326cc3abbee95df49fae3208536ee0e8110f0438e3a3"} Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.069232 4894 generic.go:334] "Generic (PLEG): container finished" podID="97355ddc-532e-46a2-a568-31f4a5e07030" containerID="cf236dedc39a8f67620519bcb9ba4e9cef45cc34cf202fb40082f8055745524a" exitCode=0 Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.069281 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97355ddc-532e-46a2-a568-31f4a5e07030","Type":"ContainerDied","Data":"cf236dedc39a8f67620519bcb9ba4e9cef45cc34cf202fb40082f8055745524a"} Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.628943 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:47 crc kubenswrapper[4894]: I1210 19:15:47.631561 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.079740 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8f81cda0-aeed-41c1-9941-5f1ba92a88c5","Type":"ContainerStarted","Data":"7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c"} Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.080248 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.082360 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97355ddc-532e-46a2-a568-31f4a5e07030","Type":"ContainerStarted","Data":"34c65a66b3db9b6eac6c97117490d462700c7f0cea9df74fa545c757a24f8484"} Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.082546 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.084693 4894 generic.go:334] "Generic (PLEG): container finished" podID="253af012-5d33-45ed-b30c-c2d114635ab3" containerID="0c743f4c3522c30eaf964cce3270510dd52fdcb74163f7f9e3730046b5c6af12" exitCode=0 Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.084751 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-twfpr" event={"ID":"253af012-5d33-45ed-b30c-c2d114635ab3","Type":"ContainerDied","Data":"0c743f4c3522c30eaf964cce3270510dd52fdcb74163f7f9e3730046b5c6af12"} Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.088539 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.105991 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.309649101 podStartE2EDuration="1m24.105972615s" podCreationTimestamp="2025-12-10 19:14:24 +0000 UTC" firstStartedPulling="2025-12-10 19:14:26.703906944 +0000 UTC m=+1149.498754736" lastFinishedPulling="2025-12-10 19:15:10.500230458 +0000 UTC m=+1193.295078250" observedRunningTime="2025-12-10 19:15:48.101979846 +0000 UTC m=+1230.896827648" watchObservedRunningTime="2025-12-10 19:15:48.105972615 +0000 UTC m=+1230.900820407" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.182069 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.204612245 podStartE2EDuration="1m24.182048886s" podCreationTimestamp="2025-12-10 19:14:24 +0000 UTC" firstStartedPulling="2025-12-10 19:14:26.307401656 +0000 UTC m=+1149.102249448" lastFinishedPulling="2025-12-10 19:15:10.284838287 +0000 UTC m=+1193.079686089" observedRunningTime="2025-12-10 19:15:48.175464806 +0000 UTC m=+1230.970312608" watchObservedRunningTime="2025-12-10 19:15:48.182048886 +0000 UTC m=+1230.976896678" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.907878 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9723-account-create-update-4b25k"] Dec 10 19:15:48 crc kubenswrapper[4894]: E1210 19:15:48.908507 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee50bb9-aa22-44b0-b596-c3f64d22c8a3" containerName="swift-ring-rebalance" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.908521 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee50bb9-aa22-44b0-b596-c3f64d22c8a3" containerName="swift-ring-rebalance" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.908754 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee50bb9-aa22-44b0-b596-c3f64d22c8a3" containerName="swift-ring-rebalance" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.909456 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.916364 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 10 19:15:48 crc kubenswrapper[4894]: I1210 19:15:48.930982 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9723-account-create-update-4b25k"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.016908 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lsxb5"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.018209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.032113 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3578f846-54cc-4fa8-8d9e-3dacbbf273db-operator-scripts\") pod \"keystone-9723-account-create-update-4b25k\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.032192 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwwk\" (UniqueName: \"kubernetes.io/projected/3578f846-54cc-4fa8-8d9e-3dacbbf273db-kube-api-access-fmwwk\") pod \"keystone-9723-account-create-update-4b25k\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.040360 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lsxb5"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.114502 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-pgjdk"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.115979 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.130367 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pgjdk"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.142000 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12383739-aa0a-4040-8d8d-f28fe4f155d0-operator-scripts\") pod \"keystone-db-create-lsxb5\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.142058 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3578f846-54cc-4fa8-8d9e-3dacbbf273db-operator-scripts\") pod \"keystone-9723-account-create-update-4b25k\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.142115 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnr7l\" (UniqueName: \"kubernetes.io/projected/12383739-aa0a-4040-8d8d-f28fe4f155d0-kube-api-access-hnr7l\") pod \"keystone-db-create-lsxb5\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.142167 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwwk\" (UniqueName: \"kubernetes.io/projected/3578f846-54cc-4fa8-8d9e-3dacbbf273db-kube-api-access-fmwwk\") pod \"keystone-9723-account-create-update-4b25k\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.144135 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3578f846-54cc-4fa8-8d9e-3dacbbf273db-operator-scripts\") pod \"keystone-9723-account-create-update-4b25k\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.183995 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwwk\" (UniqueName: \"kubernetes.io/projected/3578f846-54cc-4fa8-8d9e-3dacbbf273db-kube-api-access-fmwwk\") pod \"keystone-9723-account-create-update-4b25k\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.209746 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4c8b-account-create-update-wdpnl"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.211073 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.213600 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.233180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.245109 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12383739-aa0a-4040-8d8d-f28fe4f155d0-operator-scripts\") pod \"keystone-db-create-lsxb5\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.245199 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnr7l\" (UniqueName: \"kubernetes.io/projected/12383739-aa0a-4040-8d8d-f28fe4f155d0-kube-api-access-hnr7l\") pod \"keystone-db-create-lsxb5\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.246447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12383739-aa0a-4040-8d8d-f28fe4f155d0-operator-scripts\") pod \"keystone-db-create-lsxb5\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.256101 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4c8b-account-create-update-wdpnl"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.271552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnr7l\" (UniqueName: \"kubernetes.io/projected/12383739-aa0a-4040-8d8d-f28fe4f155d0-kube-api-access-hnr7l\") pod \"keystone-db-create-lsxb5\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.336248 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.346796 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/950c46f1-26b7-4213-9db5-e49f5cd4ae31-operator-scripts\") pod \"placement-db-create-pgjdk\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.347286 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538ed591-7bb5-4109-942e-6a5b83171db4-operator-scripts\") pod \"placement-4c8b-account-create-update-wdpnl\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.347314 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqxnt\" (UniqueName: \"kubernetes.io/projected/538ed591-7bb5-4109-942e-6a5b83171db4-kube-api-access-vqxnt\") pod \"placement-4c8b-account-create-update-wdpnl\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.347417 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j5vp\" (UniqueName: \"kubernetes.io/projected/950c46f1-26b7-4213-9db5-e49f5cd4ae31-kube-api-access-6j5vp\") pod \"placement-db-create-pgjdk\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.451960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j5vp\" (UniqueName: \"kubernetes.io/projected/950c46f1-26b7-4213-9db5-e49f5cd4ae31-kube-api-access-6j5vp\") pod \"placement-db-create-pgjdk\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.452038 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/950c46f1-26b7-4213-9db5-e49f5cd4ae31-operator-scripts\") pod \"placement-db-create-pgjdk\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.452116 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538ed591-7bb5-4109-942e-6a5b83171db4-operator-scripts\") pod \"placement-4c8b-account-create-update-wdpnl\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.452134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqxnt\" (UniqueName: \"kubernetes.io/projected/538ed591-7bb5-4109-942e-6a5b83171db4-kube-api-access-vqxnt\") pod \"placement-4c8b-account-create-update-wdpnl\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.453560 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/950c46f1-26b7-4213-9db5-e49f5cd4ae31-operator-scripts\") pod \"placement-db-create-pgjdk\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.454024 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538ed591-7bb5-4109-942e-6a5b83171db4-operator-scripts\") pod \"placement-4c8b-account-create-update-wdpnl\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.474971 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-vq968"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.476257 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.477402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqxnt\" (UniqueName: \"kubernetes.io/projected/538ed591-7bb5-4109-942e-6a5b83171db4-kube-api-access-vqxnt\") pod \"placement-4c8b-account-create-update-wdpnl\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.478142 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-vq968"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.480188 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j5vp\" (UniqueName: \"kubernetes.io/projected/950c46f1-26b7-4213-9db5-e49f5cd4ae31-kube-api-access-6j5vp\") pod \"placement-db-create-pgjdk\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.486069 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.537146 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.584369 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-d80b-account-create-update-9w4rh"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.586015 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.608460 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.623155 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d80b-account-create-update-9w4rh"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.659876 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85391f2b-d55c-4abf-a95f-562c9be55a78-operator-scripts\") pod \"glance-db-create-vq968\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.660061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cjds\" (UniqueName: \"kubernetes.io/projected/85391f2b-d55c-4abf-a95f-562c9be55a78-kube-api-access-5cjds\") pod \"glance-db-create-vq968\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.698887 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.761695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp7vn\" (UniqueName: \"kubernetes.io/projected/253af012-5d33-45ed-b30c-c2d114635ab3-kube-api-access-wp7vn\") pod \"253af012-5d33-45ed-b30c-c2d114635ab3\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.761815 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-scripts\") pod \"253af012-5d33-45ed-b30c-c2d114635ab3\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762308 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-additional-scripts\") pod \"253af012-5d33-45ed-b30c-c2d114635ab3\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762352 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run\") pod \"253af012-5d33-45ed-b30c-c2d114635ab3\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762453 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run-ovn\") pod \"253af012-5d33-45ed-b30c-c2d114635ab3\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762480 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-log-ovn\") pod \"253af012-5d33-45ed-b30c-c2d114635ab3\" (UID: \"253af012-5d33-45ed-b30c-c2d114635ab3\") " Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762732 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cjds\" (UniqueName: \"kubernetes.io/projected/85391f2b-d55c-4abf-a95f-562c9be55a78-kube-api-access-5cjds\") pod \"glance-db-create-vq968\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762724 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run" (OuterVolumeSpecName: "var-run") pod "253af012-5d33-45ed-b30c-c2d114635ab3" (UID: "253af012-5d33-45ed-b30c-c2d114635ab3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.762986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "253af012-5d33-45ed-b30c-c2d114635ab3" (UID: "253af012-5d33-45ed-b30c-c2d114635ab3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.763117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "253af012-5d33-45ed-b30c-c2d114635ab3" (UID: "253af012-5d33-45ed-b30c-c2d114635ab3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.763354 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "253af012-5d33-45ed-b30c-c2d114635ab3" (UID: "253af012-5d33-45ed-b30c-c2d114635ab3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.763399 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd2lj\" (UniqueName: \"kubernetes.io/projected/1eb80a4c-f422-4623-841b-ab19ea29f567-kube-api-access-wd2lj\") pod \"glance-d80b-account-create-update-9w4rh\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.763803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1eb80a4c-f422-4623-841b-ab19ea29f567-operator-scripts\") pod \"glance-d80b-account-create-update-9w4rh\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.764089 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85391f2b-d55c-4abf-a95f-562c9be55a78-operator-scripts\") pod \"glance-db-create-vq968\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.764171 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.764181 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.764189 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.764197 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/253af012-5d33-45ed-b30c-c2d114635ab3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.764624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-scripts" (OuterVolumeSpecName: "scripts") pod "253af012-5d33-45ed-b30c-c2d114635ab3" (UID: "253af012-5d33-45ed-b30c-c2d114635ab3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.765005 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85391f2b-d55c-4abf-a95f-562c9be55a78-operator-scripts\") pod \"glance-db-create-vq968\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.794135 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/253af012-5d33-45ed-b30c-c2d114635ab3-kube-api-access-wp7vn" (OuterVolumeSpecName: "kube-api-access-wp7vn") pod "253af012-5d33-45ed-b30c-c2d114635ab3" (UID: "253af012-5d33-45ed-b30c-c2d114635ab3"). InnerVolumeSpecName "kube-api-access-wp7vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.799248 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cjds\" (UniqueName: \"kubernetes.io/projected/85391f2b-d55c-4abf-a95f-562c9be55a78-kube-api-access-5cjds\") pod \"glance-db-create-vq968\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.831579 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vq968" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.851829 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9723-account-create-update-4b25k"] Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.865156 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd2lj\" (UniqueName: \"kubernetes.io/projected/1eb80a4c-f422-4623-841b-ab19ea29f567-kube-api-access-wd2lj\") pod \"glance-d80b-account-create-update-9w4rh\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.865219 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1eb80a4c-f422-4623-841b-ab19ea29f567-operator-scripts\") pod \"glance-d80b-account-create-update-9w4rh\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.865352 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp7vn\" (UniqueName: \"kubernetes.io/projected/253af012-5d33-45ed-b30c-c2d114635ab3-kube-api-access-wp7vn\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.865375 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/253af012-5d33-45ed-b30c-c2d114635ab3-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.866091 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1eb80a4c-f422-4623-841b-ab19ea29f567-operator-scripts\") pod \"glance-d80b-account-create-update-9w4rh\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.884774 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd2lj\" (UniqueName: \"kubernetes.io/projected/1eb80a4c-f422-4623-841b-ab19ea29f567-kube-api-access-wd2lj\") pod \"glance-d80b-account-create-update-9w4rh\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:49 crc kubenswrapper[4894]: I1210 19:15:49.989541 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.008808 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lsxb5"] Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.125107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lsxb5" event={"ID":"12383739-aa0a-4040-8d8d-f28fe4f155d0","Type":"ContainerStarted","Data":"f0be6255da00b10a4de0de9bcdee1b707d2adf1c0fc04d971b37a9e2a69c5a0a"} Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.127170 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9723-account-create-update-4b25k" event={"ID":"3578f846-54cc-4fa8-8d9e-3dacbbf273db","Type":"ContainerStarted","Data":"a4a8060b1c0e3b910adcc3bdab022f87796f5d1a9ae6916e69bac93e36a0c3d2"} Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.127195 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9723-account-create-update-4b25k" event={"ID":"3578f846-54cc-4fa8-8d9e-3dacbbf273db","Type":"ContainerStarted","Data":"169470f2f26c2257e9606dbd75acd68364bc1b7fa99f3e543e234944f4156046"} Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.132402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-twfpr" event={"ID":"253af012-5d33-45ed-b30c-c2d114635ab3","Type":"ContainerDied","Data":"ea1af466c8aea196c014326cc3abbee95df49fae3208536ee0e8110f0438e3a3"} Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.132440 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea1af466c8aea196c014326cc3abbee95df49fae3208536ee0e8110f0438e3a3" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.132471 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-twfpr" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.175666 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9723-account-create-update-4b25k" podStartSLOduration=2.175635434 podStartE2EDuration="2.175635434s" podCreationTimestamp="2025-12-10 19:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:50.143387036 +0000 UTC m=+1232.938234838" watchObservedRunningTime="2025-12-10 19:15:50.175635434 +0000 UTC m=+1232.970483226" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.200649 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4c8b-account-create-update-wdpnl"] Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.241593 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pgjdk"] Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.346151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-vq968"] Dec 10 19:15:50 crc kubenswrapper[4894]: W1210 19:15:50.695442 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb80a4c_f422_4623_841b_ab19ea29f567.slice/crio-2d52e87217926b2cd5d1e4e4339833b10bd4ea10fcfa2b7677ca294b249d0e70 WatchSource:0}: Error finding container 2d52e87217926b2cd5d1e4e4339833b10bd4ea10fcfa2b7677ca294b249d0e70: Status 404 returned error can't find the container with id 2d52e87217926b2cd5d1e4e4339833b10bd4ea10fcfa2b7677ca294b249d0e70 Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.698411 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d80b-account-create-update-9w4rh"] Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.863694 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-h2552-config-twfpr"] Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.870951 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-h2552-config-twfpr"] Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.938552 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-h2552" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.960706 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-h2552-config-2gqtw"] Dec 10 19:15:50 crc kubenswrapper[4894]: E1210 19:15:50.962624 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253af012-5d33-45ed-b30c-c2d114635ab3" containerName="ovn-config" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.962653 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="253af012-5d33-45ed-b30c-c2d114635ab3" containerName="ovn-config" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.962902 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="253af012-5d33-45ed-b30c-c2d114635ab3" containerName="ovn-config" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.963766 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.967025 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 19:15:50 crc kubenswrapper[4894]: I1210 19:15:50.969606 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h2552-config-2gqtw"] Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.003349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-log-ovn\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.003413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run-ovn\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.003501 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.003524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-additional-scripts\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.003578 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6nc4\" (UniqueName: \"kubernetes.io/projected/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-kube-api-access-h6nc4\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.003669 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-scripts\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.105065 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.105461 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-additional-scripts\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.105414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.106393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-additional-scripts\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.106493 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6nc4\" (UniqueName: \"kubernetes.io/projected/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-kube-api-access-h6nc4\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.106900 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-scripts\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.108801 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-scripts\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.108954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-log-ovn\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.108994 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run-ovn\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.109057 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-log-ovn\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.109154 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run-ovn\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.132482 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6nc4\" (UniqueName: \"kubernetes.io/projected/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-kube-api-access-h6nc4\") pod \"ovn-controller-h2552-config-2gqtw\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.140660 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.142763 4894 generic.go:334] "Generic (PLEG): container finished" podID="3578f846-54cc-4fa8-8d9e-3dacbbf273db" containerID="a4a8060b1c0e3b910adcc3bdab022f87796f5d1a9ae6916e69bac93e36a0c3d2" exitCode=0 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.142829 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9723-account-create-update-4b25k" event={"ID":"3578f846-54cc-4fa8-8d9e-3dacbbf273db","Type":"ContainerDied","Data":"a4a8060b1c0e3b910adcc3bdab022f87796f5d1a9ae6916e69bac93e36a0c3d2"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.149235 4894 generic.go:334] "Generic (PLEG): container finished" podID="950c46f1-26b7-4213-9db5-e49f5cd4ae31" containerID="2c23663e860f29555de487eb4deecf1f0e5773a2613775ee2661cc4a4be06f01" exitCode=0 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.149312 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pgjdk" event={"ID":"950c46f1-26b7-4213-9db5-e49f5cd4ae31","Type":"ContainerDied","Data":"2c23663e860f29555de487eb4deecf1f0e5773a2613775ee2661cc4a4be06f01"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.149343 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pgjdk" event={"ID":"950c46f1-26b7-4213-9db5-e49f5cd4ae31","Type":"ContainerStarted","Data":"ddd3e839ee72f2f872d665f10b8e4dedac728d96431e436b840366f5d9f06c5a"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.151764 4894 generic.go:334] "Generic (PLEG): container finished" podID="538ed591-7bb5-4109-942e-6a5b83171db4" containerID="7559549a1f572e5280f008033089971297e58264d079106ec602858ee9371b64" exitCode=0 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.151887 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c8b-account-create-update-wdpnl" event={"ID":"538ed591-7bb5-4109-942e-6a5b83171db4","Type":"ContainerDied","Data":"7559549a1f572e5280f008033089971297e58264d079106ec602858ee9371b64"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.151913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c8b-account-create-update-wdpnl" event={"ID":"538ed591-7bb5-4109-942e-6a5b83171db4","Type":"ContainerStarted","Data":"418a5d9b60b5ec531d26bf950448bdb79e84cb9011175a43affaf2d357ccd89c"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.158120 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d80b-account-create-update-9w4rh" event={"ID":"1eb80a4c-f422-4623-841b-ab19ea29f567","Type":"ContainerStarted","Data":"a6d372f049bdc761d0128659b5797103055988677c75f7f760bf046e5bee592c"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.158162 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d80b-account-create-update-9w4rh" event={"ID":"1eb80a4c-f422-4623-841b-ab19ea29f567","Type":"ContainerStarted","Data":"2d52e87217926b2cd5d1e4e4339833b10bd4ea10fcfa2b7677ca294b249d0e70"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.160441 4894 generic.go:334] "Generic (PLEG): container finished" podID="12383739-aa0a-4040-8d8d-f28fe4f155d0" containerID="cf21f959697e074ec576f1d07143a2c73ede2db9611fc61d3cace6b9b0b82e92" exitCode=0 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.160511 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lsxb5" event={"ID":"12383739-aa0a-4040-8d8d-f28fe4f155d0","Type":"ContainerDied","Data":"cf21f959697e074ec576f1d07143a2c73ede2db9611fc61d3cace6b9b0b82e92"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.161941 4894 generic.go:334] "Generic (PLEG): container finished" podID="85391f2b-d55c-4abf-a95f-562c9be55a78" containerID="4fa854ec517d77864924b4341c079a285b769a523ab672b091a0676d8391da5a" exitCode=0 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.162002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-vq968" event={"ID":"85391f2b-d55c-4abf-a95f-562c9be55a78","Type":"ContainerDied","Data":"4fa854ec517d77864924b4341c079a285b769a523ab672b091a0676d8391da5a"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.162026 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-vq968" event={"ID":"85391f2b-d55c-4abf-a95f-562c9be55a78","Type":"ContainerStarted","Data":"3e556dec768d27252388eef17f3afacefcb1214cc02739e2c35368b21a4b11b0"} Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.554623 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="253af012-5d33-45ed-b30c-c2d114635ab3" path="/var/lib/kubelet/pods/253af012-5d33-45ed-b30c-c2d114635ab3/volumes" Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.610370 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-h2552-config-2gqtw"] Dec 10 19:15:51 crc kubenswrapper[4894]: W1210 19:15:51.615767 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e22e3c4_ba6e_41f7_ac08_41cfe5ff4af1.slice/crio-81b4bac4460029f77a72e5e9316dca7feefe3b68c275cd4dfe972e5f940320e6 WatchSource:0}: Error finding container 81b4bac4460029f77a72e5e9316dca7feefe3b68c275cd4dfe972e5f940320e6: Status 404 returned error can't find the container with id 81b4bac4460029f77a72e5e9316dca7feefe3b68c275cd4dfe972e5f940320e6 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.950520 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.951092 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="prometheus" containerID="cri-o://ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec" gracePeriod=600 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.951207 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="config-reloader" containerID="cri-o://d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5" gracePeriod=600 Dec 10 19:15:51 crc kubenswrapper[4894]: I1210 19:15:51.951403 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="thanos-sidecar" containerID="cri-o://4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602" gracePeriod=600 Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.172181 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-2gqtw" event={"ID":"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1","Type":"ContainerStarted","Data":"0cb0a7113b5b12e095db65306bf061698d8600b4438f43afad3e4c4751eaf9a2"} Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.172239 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-2gqtw" event={"ID":"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1","Type":"ContainerStarted","Data":"81b4bac4460029f77a72e5e9316dca7feefe3b68c275cd4dfe972e5f940320e6"} Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.174196 4894 generic.go:334] "Generic (PLEG): container finished" podID="1eb80a4c-f422-4623-841b-ab19ea29f567" containerID="a6d372f049bdc761d0128659b5797103055988677c75f7f760bf046e5bee592c" exitCode=0 Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.174238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d80b-account-create-update-9w4rh" event={"ID":"1eb80a4c-f422-4623-841b-ab19ea29f567","Type":"ContainerDied","Data":"a6d372f049bdc761d0128659b5797103055988677c75f7f760bf046e5bee592c"} Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.179981 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerID="4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602" exitCode=0 Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.180015 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerID="ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec" exitCode=0 Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.180175 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerDied","Data":"4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602"} Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.180203 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerDied","Data":"ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec"} Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.200529 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-h2552-config-2gqtw" podStartSLOduration=2.200508049 podStartE2EDuration="2.200508049s" podCreationTimestamp="2025-12-10 19:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:15:52.191082228 +0000 UTC m=+1234.985930020" watchObservedRunningTime="2025-12-10 19:15:52.200508049 +0000 UTC m=+1234.995355861" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.627717 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.628910 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.113:9090/-/ready\": dial tcp 10.217.0.113:9090: connect: connection refused" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.744450 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3578f846-54cc-4fa8-8d9e-3dacbbf273db-operator-scripts\") pod \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.744625 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmwwk\" (UniqueName: \"kubernetes.io/projected/3578f846-54cc-4fa8-8d9e-3dacbbf273db-kube-api-access-fmwwk\") pod \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\" (UID: \"3578f846-54cc-4fa8-8d9e-3dacbbf273db\") " Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.746321 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3578f846-54cc-4fa8-8d9e-3dacbbf273db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3578f846-54cc-4fa8-8d9e-3dacbbf273db" (UID: "3578f846-54cc-4fa8-8d9e-3dacbbf273db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.775036 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3578f846-54cc-4fa8-8d9e-3dacbbf273db-kube-api-access-fmwwk" (OuterVolumeSpecName: "kube-api-access-fmwwk") pod "3578f846-54cc-4fa8-8d9e-3dacbbf273db" (UID: "3578f846-54cc-4fa8-8d9e-3dacbbf273db"). InnerVolumeSpecName "kube-api-access-fmwwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.848520 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmwwk\" (UniqueName: \"kubernetes.io/projected/3578f846-54cc-4fa8-8d9e-3dacbbf273db-kube-api-access-fmwwk\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.848580 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3578f846-54cc-4fa8-8d9e-3dacbbf273db-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.996141 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vq968" Dec 10 19:15:52 crc kubenswrapper[4894]: I1210 19:15:52.996789 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.001940 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.018238 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053406 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j5vp\" (UniqueName: \"kubernetes.io/projected/950c46f1-26b7-4213-9db5-e49f5cd4ae31-kube-api-access-6j5vp\") pod \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053495 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnr7l\" (UniqueName: \"kubernetes.io/projected/12383739-aa0a-4040-8d8d-f28fe4f155d0-kube-api-access-hnr7l\") pod \"12383739-aa0a-4040-8d8d-f28fe4f155d0\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053582 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12383739-aa0a-4040-8d8d-f28fe4f155d0-operator-scripts\") pod \"12383739-aa0a-4040-8d8d-f28fe4f155d0\" (UID: \"12383739-aa0a-4040-8d8d-f28fe4f155d0\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053631 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqxnt\" (UniqueName: \"kubernetes.io/projected/538ed591-7bb5-4109-942e-6a5b83171db4-kube-api-access-vqxnt\") pod \"538ed591-7bb5-4109-942e-6a5b83171db4\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053647 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85391f2b-d55c-4abf-a95f-562c9be55a78-operator-scripts\") pod \"85391f2b-d55c-4abf-a95f-562c9be55a78\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053670 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/950c46f1-26b7-4213-9db5-e49f5cd4ae31-operator-scripts\") pod \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\" (UID: \"950c46f1-26b7-4213-9db5-e49f5cd4ae31\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053711 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cjds\" (UniqueName: \"kubernetes.io/projected/85391f2b-d55c-4abf-a95f-562c9be55a78-kube-api-access-5cjds\") pod \"85391f2b-d55c-4abf-a95f-562c9be55a78\" (UID: \"85391f2b-d55c-4abf-a95f-562c9be55a78\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.053736 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538ed591-7bb5-4109-942e-6a5b83171db4-operator-scripts\") pod \"538ed591-7bb5-4109-942e-6a5b83171db4\" (UID: \"538ed591-7bb5-4109-942e-6a5b83171db4\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.054559 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/538ed591-7bb5-4109-942e-6a5b83171db4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "538ed591-7bb5-4109-942e-6a5b83171db4" (UID: "538ed591-7bb5-4109-942e-6a5b83171db4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.055560 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950c46f1-26b7-4213-9db5-e49f5cd4ae31-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "950c46f1-26b7-4213-9db5-e49f5cd4ae31" (UID: "950c46f1-26b7-4213-9db5-e49f5cd4ae31"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.055919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85391f2b-d55c-4abf-a95f-562c9be55a78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85391f2b-d55c-4abf-a95f-562c9be55a78" (UID: "85391f2b-d55c-4abf-a95f-562c9be55a78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.059486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12383739-aa0a-4040-8d8d-f28fe4f155d0-kube-api-access-hnr7l" (OuterVolumeSpecName: "kube-api-access-hnr7l") pod "12383739-aa0a-4040-8d8d-f28fe4f155d0" (UID: "12383739-aa0a-4040-8d8d-f28fe4f155d0"). InnerVolumeSpecName "kube-api-access-hnr7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.059587 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/950c46f1-26b7-4213-9db5-e49f5cd4ae31-kube-api-access-6j5vp" (OuterVolumeSpecName: "kube-api-access-6j5vp") pod "950c46f1-26b7-4213-9db5-e49f5cd4ae31" (UID: "950c46f1-26b7-4213-9db5-e49f5cd4ae31"). InnerVolumeSpecName "kube-api-access-6j5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.060493 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538ed591-7bb5-4109-942e-6a5b83171db4-kube-api-access-vqxnt" (OuterVolumeSpecName: "kube-api-access-vqxnt") pod "538ed591-7bb5-4109-942e-6a5b83171db4" (UID: "538ed591-7bb5-4109-942e-6a5b83171db4"). InnerVolumeSpecName "kube-api-access-vqxnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.062006 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85391f2b-d55c-4abf-a95f-562c9be55a78-kube-api-access-5cjds" (OuterVolumeSpecName: "kube-api-access-5cjds") pod "85391f2b-d55c-4abf-a95f-562c9be55a78" (UID: "85391f2b-d55c-4abf-a95f-562c9be55a78"). InnerVolumeSpecName "kube-api-access-5cjds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.062071 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12383739-aa0a-4040-8d8d-f28fe4f155d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12383739-aa0a-4040-8d8d-f28fe4f155d0" (UID: "12383739-aa0a-4040-8d8d-f28fe4f155d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.129818 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155731 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12383739-aa0a-4040-8d8d-f28fe4f155d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155769 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqxnt\" (UniqueName: \"kubernetes.io/projected/538ed591-7bb5-4109-942e-6a5b83171db4-kube-api-access-vqxnt\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155783 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85391f2b-d55c-4abf-a95f-562c9be55a78-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155794 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/950c46f1-26b7-4213-9db5-e49f5cd4ae31-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155805 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cjds\" (UniqueName: \"kubernetes.io/projected/85391f2b-d55c-4abf-a95f-562c9be55a78-kube-api-access-5cjds\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155814 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/538ed591-7bb5-4109-942e-6a5b83171db4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155822 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j5vp\" (UniqueName: \"kubernetes.io/projected/950c46f1-26b7-4213-9db5-e49f5cd4ae31-kube-api-access-6j5vp\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.155831 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnr7l\" (UniqueName: \"kubernetes.io/projected/12383739-aa0a-4040-8d8d-f28fe4f155d0-kube-api-access-hnr7l\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.195019 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.198230 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerID="d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5" exitCode=0 Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.198279 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerDied","Data":"d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.198711 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2f3a6b99-5cd8-4149-8825-e19613a13135","Type":"ContainerDied","Data":"5edc5921c54b5bf44ceb2939bf610e5b1771551145fd39870376bd643a6bcf32"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.198738 4894 scope.go:117] "RemoveContainer" containerID="4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.202451 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-vq968" event={"ID":"85391f2b-d55c-4abf-a95f-562c9be55a78","Type":"ContainerDied","Data":"3e556dec768d27252388eef17f3afacefcb1214cc02739e2c35368b21a4b11b0"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.202482 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e556dec768d27252388eef17f3afacefcb1214cc02739e2c35368b21a4b11b0" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.202535 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vq968" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.205142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9723-account-create-update-4b25k" event={"ID":"3578f846-54cc-4fa8-8d9e-3dacbbf273db","Type":"ContainerDied","Data":"169470f2f26c2257e9606dbd75acd68364bc1b7fa99f3e543e234944f4156046"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.205178 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="169470f2f26c2257e9606dbd75acd68364bc1b7fa99f3e543e234944f4156046" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.205256 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9723-account-create-update-4b25k" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.208631 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pgjdk" event={"ID":"950c46f1-26b7-4213-9db5-e49f5cd4ae31","Type":"ContainerDied","Data":"ddd3e839ee72f2f872d665f10b8e4dedac728d96431e436b840366f5d9f06c5a"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.208675 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddd3e839ee72f2f872d665f10b8e4dedac728d96431e436b840366f5d9f06c5a" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.208735 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pgjdk" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.217317 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" containerID="0cb0a7113b5b12e095db65306bf061698d8600b4438f43afad3e4c4751eaf9a2" exitCode=0 Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.217371 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-2gqtw" event={"ID":"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1","Type":"ContainerDied","Data":"0cb0a7113b5b12e095db65306bf061698d8600b4438f43afad3e4c4751eaf9a2"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.226276 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c8b-account-create-update-wdpnl" event={"ID":"538ed591-7bb5-4109-942e-6a5b83171db4","Type":"ContainerDied","Data":"418a5d9b60b5ec531d26bf950448bdb79e84cb9011175a43affaf2d357ccd89c"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.226305 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="418a5d9b60b5ec531d26bf950448bdb79e84cb9011175a43affaf2d357ccd89c" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.226367 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c8b-account-create-update-wdpnl" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.241468 4894 scope.go:117] "RemoveContainer" containerID="d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.243096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d80b-account-create-update-9w4rh" event={"ID":"1eb80a4c-f422-4623-841b-ab19ea29f567","Type":"ContainerDied","Data":"2d52e87217926b2cd5d1e4e4339833b10bd4ea10fcfa2b7677ca294b249d0e70"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.243132 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d52e87217926b2cd5d1e4e4339833b10bd4ea10fcfa2b7677ca294b249d0e70" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.243316 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d80b-account-create-update-9w4rh" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.252299 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lsxb5" event={"ID":"12383739-aa0a-4040-8d8d-f28fe4f155d0","Type":"ContainerDied","Data":"f0be6255da00b10a4de0de9bcdee1b707d2adf1c0fc04d971b37a9e2a69c5a0a"} Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.252340 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0be6255da00b10a4de0de9bcdee1b707d2adf1c0fc04d971b37a9e2a69c5a0a" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.252411 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lsxb5" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.257135 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-web-config\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.259921 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-config\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260021 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-tls-assets\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd2lj\" (UniqueName: \"kubernetes.io/projected/1eb80a4c-f422-4623-841b-ab19ea29f567-kube-api-access-wd2lj\") pod \"1eb80a4c-f422-4623-841b-ab19ea29f567\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260097 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t2lj\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-kube-api-access-6t2lj\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260135 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f3a6b99-5cd8-4149-8825-e19613a13135-config-out\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260246 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260289 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-thanos-prometheus-http-client-file\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260354 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f3a6b99-5cd8-4149-8825-e19613a13135-prometheus-metric-storage-rulefiles-0\") pod \"2f3a6b99-5cd8-4149-8825-e19613a13135\" (UID: \"2f3a6b99-5cd8-4149-8825-e19613a13135\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260380 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1eb80a4c-f422-4623-841b-ab19ea29f567-operator-scripts\") pod \"1eb80a4c-f422-4623-841b-ab19ea29f567\" (UID: \"1eb80a4c-f422-4623-841b-ab19ea29f567\") " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260378 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-config" (OuterVolumeSpecName: "config") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.260869 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.261524 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f3a6b99-5cd8-4149-8825-e19613a13135-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.261863 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb80a4c-f422-4623-841b-ab19ea29f567-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1eb80a4c-f422-4623-841b-ab19ea29f567" (UID: "1eb80a4c-f422-4623-841b-ab19ea29f567"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.264293 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb80a4c-f422-4623-841b-ab19ea29f567-kube-api-access-wd2lj" (OuterVolumeSpecName: "kube-api-access-wd2lj") pod "1eb80a4c-f422-4623-841b-ab19ea29f567" (UID: "1eb80a4c-f422-4623-841b-ab19ea29f567"). InnerVolumeSpecName "kube-api-access-wd2lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.266275 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.268157 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.269089 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3a6b99-5cd8-4149-8825-e19613a13135-config-out" (OuterVolumeSpecName: "config-out") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.272162 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-kube-api-access-6t2lj" (OuterVolumeSpecName: "kube-api-access-6t2lj") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "kube-api-access-6t2lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.295041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "pvc-14b10d10-705e-47b1-9407-51d03a71541c". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.302509 4894 scope.go:117] "RemoveContainer" containerID="ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.316294 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-web-config" (OuterVolumeSpecName: "web-config") pod "2f3a6b99-5cd8-4149-8825-e19613a13135" (UID: "2f3a6b99-5cd8-4149-8825-e19613a13135"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.337635 4894 scope.go:117] "RemoveContainer" containerID="08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.361816 4894 scope.go:117] "RemoveContainer" containerID="4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602" Dec 10 19:15:53 crc kubenswrapper[4894]: E1210 19:15:53.362181 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602\": container with ID starting with 4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602 not found: ID does not exist" containerID="4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362209 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602"} err="failed to get container status \"4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602\": rpc error: code = NotFound desc = could not find container \"4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602\": container with ID starting with 4b52a0b6012ea87751e636df1dcda995342c636a1741b0474e730f74e4589602 not found: ID does not exist" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362227 4894 scope.go:117] "RemoveContainer" containerID="d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5" Dec 10 19:15:53 crc kubenswrapper[4894]: E1210 19:15:53.362502 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5\": container with ID starting with d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5 not found: ID does not exist" containerID="d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362521 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5"} err="failed to get container status \"d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5\": rpc error: code = NotFound desc = could not find container \"d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5\": container with ID starting with d7a26e9db53558268d408d65d21bcb3b6be71c9813d6b13573a9488bcbf876b5 not found: ID does not exist" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362532 4894 scope.go:117] "RemoveContainer" containerID="ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec" Dec 10 19:15:53 crc kubenswrapper[4894]: E1210 19:15:53.362748 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec\": container with ID starting with ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec not found: ID does not exist" containerID="ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362766 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec"} err="failed to get container status \"ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec\": rpc error: code = NotFound desc = could not find container \"ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec\": container with ID starting with ed8ec3bd729712e3a4e02c32aaec42852d335dd0c2cd81429f0e875de9bd0aec not found: ID does not exist" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362777 4894 scope.go:117] "RemoveContainer" containerID="08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f" Dec 10 19:15:53 crc kubenswrapper[4894]: E1210 19:15:53.362975 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f\": container with ID starting with 08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f not found: ID does not exist" containerID="08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.362995 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f"} err="failed to get container status \"08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f\": rpc error: code = NotFound desc = could not find container \"08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f\": container with ID starting with 08516c97185646f63ceee3491fa7b33efeac10c0d31421786a937d5c8269c63f not found: ID does not exist" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363615 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd2lj\" (UniqueName: \"kubernetes.io/projected/1eb80a4c-f422-4623-841b-ab19ea29f567-kube-api-access-wd2lj\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363637 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t2lj\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-kube-api-access-6t2lj\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363646 4894 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2f3a6b99-5cd8-4149-8825-e19613a13135-config-out\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363683 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") on node \"crc\" " Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363696 4894 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363707 4894 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2f3a6b99-5cd8-4149-8825-e19613a13135-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363716 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1eb80a4c-f422-4623-841b-ab19ea29f567-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363725 4894 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2f3a6b99-5cd8-4149-8825-e19613a13135-web-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.363733 4894 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2f3a6b99-5cd8-4149-8825-e19613a13135-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.390413 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.390582 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-14b10d10-705e-47b1-9407-51d03a71541c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c") on node "crc" Dec 10 19:15:53 crc kubenswrapper[4894]: I1210 19:15:53.465633 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.261361 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.280368 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.291837 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5eac36f4-3cbc-48c1-9c6a-4e79de593d04-etc-swift\") pod \"swift-storage-0\" (UID: \"5eac36f4-3cbc-48c1-9c6a-4e79de593d04\") " pod="openstack/swift-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.295527 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.306890 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.339827 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340231 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85391f2b-d55c-4abf-a95f-562c9be55a78" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340251 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="85391f2b-d55c-4abf-a95f-562c9be55a78" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340265 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="thanos-sidecar" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340272 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="thanos-sidecar" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340283 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="init-config-reloader" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340289 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="init-config-reloader" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340301 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="950c46f1-26b7-4213-9db5-e49f5cd4ae31" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340308 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="950c46f1-26b7-4213-9db5-e49f5cd4ae31" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340320 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb80a4c-f422-4623-841b-ab19ea29f567" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340326 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb80a4c-f422-4623-841b-ab19ea29f567" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340344 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538ed591-7bb5-4109-942e-6a5b83171db4" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340350 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="538ed591-7bb5-4109-942e-6a5b83171db4" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340380 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3578f846-54cc-4fa8-8d9e-3dacbbf273db" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340387 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3578f846-54cc-4fa8-8d9e-3dacbbf273db" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340396 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="config-reloader" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340402 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="config-reloader" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340416 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12383739-aa0a-4040-8d8d-f28fe4f155d0" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340422 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="12383739-aa0a-4040-8d8d-f28fe4f155d0" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: E1210 19:15:54.340431 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="prometheus" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340436 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="prometheus" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340591 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="950c46f1-26b7-4213-9db5-e49f5cd4ae31" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340604 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="config-reloader" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340613 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="thanos-sidecar" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340620 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb80a4c-f422-4623-841b-ab19ea29f567" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340633 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="538ed591-7bb5-4109-942e-6a5b83171db4" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340641 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" containerName="prometheus" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340651 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="12383739-aa0a-4040-8d8d-f28fe4f155d0" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340660 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3578f846-54cc-4fa8-8d9e-3dacbbf273db" containerName="mariadb-account-create-update" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.340671 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="85391f2b-d55c-4abf-a95f-562c9be55a78" containerName="mariadb-database-create" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.342470 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.346473 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.351621 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.351621 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.351939 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-d22s9" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.351692 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.352374 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.362492 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.366482 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.454422 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491424 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491522 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdp82\" (UniqueName: \"kubernetes.io/projected/dae05eff-e7b1-460a-a374-74e698d8f0c9-kube-api-access-qdp82\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491594 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-config\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491657 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491678 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491756 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dae05eff-e7b1-460a-a374-74e698d8f0c9-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491783 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dae05eff-e7b1-460a-a374-74e698d8f0c9-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491873 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.491895 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.492182 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.492336 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dae05eff-e7b1-460a-a374-74e698d8f0c9-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.594083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dae05eff-e7b1-460a-a374-74e698d8f0c9-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.594348 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dae05eff-e7b1-460a-a374-74e698d8f0c9-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.594376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.594396 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.594953 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595002 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dae05eff-e7b1-460a-a374-74e698d8f0c9-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595025 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dae05eff-e7b1-460a-a374-74e698d8f0c9-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdp82\" (UniqueName: \"kubernetes.io/projected/dae05eff-e7b1-460a-a374-74e698d8f0c9-kube-api-access-qdp82\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-config\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595119 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.595137 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.600307 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.600868 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.600907 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ee4633b447c807486101aee2fd0379a07b7444d2812f770a7ff8debe13b85c97/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.601229 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.602694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dae05eff-e7b1-460a-a374-74e698d8f0c9-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.603265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dae05eff-e7b1-460a-a374-74e698d8f0c9-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.605394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.607923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-config\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.608073 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.612805 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae05eff-e7b1-460a-a374-74e698d8f0c9-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.622540 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdp82\" (UniqueName: \"kubernetes.io/projected/dae05eff-e7b1-460a-a374-74e698d8f0c9-kube-api-access-qdp82\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.661368 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-14b10d10-705e-47b1-9407-51d03a71541c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-14b10d10-705e-47b1-9407-51d03a71541c\") pod \"prometheus-metric-storage-0\" (UID: \"dae05eff-e7b1-460a-a374-74e698d8f0c9\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.752871 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-kxf26"] Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.754335 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.755571 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.756383 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-r2c4m" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.757866 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.765390 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kxf26"] Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.801672 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-log-ovn\") pod \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.801922 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" (UID: "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.802132 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-scripts\") pod \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.802256 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6nc4\" (UniqueName: \"kubernetes.io/projected/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-kube-api-access-h6nc4\") pod \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.802352 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-additional-scripts\") pod \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.802492 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run-ovn\") pod \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.802583 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run\") pod \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\" (UID: \"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1\") " Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803120 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbk2r\" (UniqueName: \"kubernetes.io/projected/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-kube-api-access-rbk2r\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803332 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-db-sync-config-data\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.802707 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run" (OuterVolumeSpecName: "var-run") pod "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" (UID: "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803038 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" (UID: "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803076 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" (UID: "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803608 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-config-data\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803692 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-combined-ca-bundle\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803800 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.803884 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.804023 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.804172 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.804115 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-scripts" (OuterVolumeSpecName: "scripts") pod "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" (UID: "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.807521 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-kube-api-access-h6nc4" (OuterVolumeSpecName: "kube-api-access-h6nc4") pod "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" (UID: "9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1"). InnerVolumeSpecName "kube-api-access-h6nc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.905954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbk2r\" (UniqueName: \"kubernetes.io/projected/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-kube-api-access-rbk2r\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.906636 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-db-sync-config-data\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.906675 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-config-data\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.907333 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-combined-ca-bundle\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.907379 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.907390 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6nc4\" (UniqueName: \"kubernetes.io/projected/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1-kube-api-access-h6nc4\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.910258 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-config-data\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.910316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-db-sync-config-data\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.910788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-combined-ca-bundle\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.927359 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbk2r\" (UniqueName: \"kubernetes.io/projected/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-kube-api-access-rbk2r\") pod \"glance-db-sync-kxf26\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:54 crc kubenswrapper[4894]: I1210 19:15:54.961177 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.073365 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kxf26" Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.289935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-h2552-config-2gqtw" event={"ID":"9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1","Type":"ContainerDied","Data":"81b4bac4460029f77a72e5e9316dca7feefe3b68c275cd4dfe972e5f940320e6"} Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.289977 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81b4bac4460029f77a72e5e9316dca7feefe3b68c275cd4dfe972e5f940320e6" Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.290045 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-h2552-config-2gqtw" Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.313013 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.340562 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-h2552-config-2gqtw"] Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.357961 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-h2552-config-2gqtw"] Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.401305 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="da36895a-2989-4110-88a2-6bf0eb5d2f5d" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.454229 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:15:55 crc kubenswrapper[4894]: W1210 19:15:55.461982 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddae05eff_e7b1_460a_a374_74e698d8f0c9.slice/crio-39153571df61b1eccfb4116832837732948ce03d6e1feba7dc9e759662c71b6e WatchSource:0}: Error finding container 39153571df61b1eccfb4116832837732948ce03d6e1feba7dc9e759662c71b6e: Status 404 returned error can't find the container with id 39153571df61b1eccfb4116832837732948ce03d6e1feba7dc9e759662c71b6e Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.554412 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f3a6b99-5cd8-4149-8825-e19613a13135" path="/var/lib/kubelet/pods/2f3a6b99-5cd8-4149-8825-e19613a13135/volumes" Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.555398 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" path="/var/lib/kubelet/pods/9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1/volumes" Dec 10 19:15:55 crc kubenswrapper[4894]: W1210 19:15:55.807037 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f1bbcc0_1d98_452d_9b61_9caf78395cb4.slice/crio-d672c0a49a4f0c2650f1195399654d8af3e32a4cb5f2d8fed97321d6f3c79880 WatchSource:0}: Error finding container d672c0a49a4f0c2650f1195399654d8af3e32a4cb5f2d8fed97321d6f3c79880: Status 404 returned error can't find the container with id d672c0a49a4f0c2650f1195399654d8af3e32a4cb5f2d8fed97321d6f3c79880 Dec 10 19:15:55 crc kubenswrapper[4894]: I1210 19:15:55.808298 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kxf26"] Dec 10 19:15:56 crc kubenswrapper[4894]: I1210 19:15:56.299191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kxf26" event={"ID":"8f1bbcc0-1d98-452d-9b61-9caf78395cb4","Type":"ContainerStarted","Data":"d672c0a49a4f0c2650f1195399654d8af3e32a4cb5f2d8fed97321d6f3c79880"} Dec 10 19:15:56 crc kubenswrapper[4894]: I1210 19:15:56.300219 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dae05eff-e7b1-460a-a374-74e698d8f0c9","Type":"ContainerStarted","Data":"39153571df61b1eccfb4116832837732948ce03d6e1feba7dc9e759662c71b6e"} Dec 10 19:15:56 crc kubenswrapper[4894]: I1210 19:15:56.311617 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"5a0fbcab6238ee13f2058c4b02094a2ab5c3307a694aa9bcc29ef8b43ae6eef0"} Dec 10 19:15:58 crc kubenswrapper[4894]: I1210 19:15:58.332870 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dae05eff-e7b1-460a-a374-74e698d8f0c9","Type":"ContainerStarted","Data":"18d9d410e25f2685e22479ca21557f3478a5d36711412bf07cb7eb024942fad1"} Dec 10 19:15:59 crc kubenswrapper[4894]: I1210 19:15:59.342099 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"80a98049e3c0a062e01e266ceccecd497f69e94dc2457f8d387a7d3b65e2f6b8"} Dec 10 19:15:59 crc kubenswrapper[4894]: I1210 19:15:59.342494 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"1a416f319e81bf38246d6034cf00b22701962b2fda3fa4bb58fea7725113a5b3"} Dec 10 19:16:00 crc kubenswrapper[4894]: I1210 19:16:00.357205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"c490f7b6e594fb2ba33f39ecd17440b8d04562147d506b42f26bcc7f462fe615"} Dec 10 19:16:00 crc kubenswrapper[4894]: I1210 19:16:00.358320 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"a30a03351a62fa7f223845d6431c0401f75c2d6c7f54622ba7ffff5a0c1ef4bb"} Dec 10 19:16:05 crc kubenswrapper[4894]: I1210 19:16:05.403297 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 10 19:16:05 crc kubenswrapper[4894]: I1210 19:16:05.408119 4894 generic.go:334] "Generic (PLEG): container finished" podID="dae05eff-e7b1-460a-a374-74e698d8f0c9" containerID="18d9d410e25f2685e22479ca21557f3478a5d36711412bf07cb7eb024942fad1" exitCode=0 Dec 10 19:16:05 crc kubenswrapper[4894]: I1210 19:16:05.408161 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dae05eff-e7b1-460a-a374-74e698d8f0c9","Type":"ContainerDied","Data":"18d9d410e25f2685e22479ca21557f3478a5d36711412bf07cb7eb024942fad1"} Dec 10 19:16:05 crc kubenswrapper[4894]: I1210 19:16:05.826767 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.107833 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.176655 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-s5x8c"] Dec 10 19:16:06 crc kubenswrapper[4894]: E1210 19:16:06.177040 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" containerName="ovn-config" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.177057 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" containerName="ovn-config" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.177251 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e22e3c4-ba6e-41f7-ac08-41cfe5ff4af1" containerName="ovn-config" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.177873 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.219055 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-s5x8c"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.270666 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87pgd\" (UniqueName: \"kubernetes.io/projected/d500410c-9685-4529-b187-750ea33d7b99-kube-api-access-87pgd\") pod \"barbican-db-create-s5x8c\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.270807 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d500410c-9685-4529-b187-750ea33d7b99-operator-scripts\") pod \"barbican-db-create-s5x8c\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.371955 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-58cb-account-create-update-n2g6b"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.373295 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.373640 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87pgd\" (UniqueName: \"kubernetes.io/projected/d500410c-9685-4529-b187-750ea33d7b99-kube-api-access-87pgd\") pod \"barbican-db-create-s5x8c\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.373768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d500410c-9685-4529-b187-750ea33d7b99-operator-scripts\") pod \"barbican-db-create-s5x8c\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.376116 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d500410c-9685-4529-b187-750ea33d7b99-operator-scripts\") pod \"barbican-db-create-s5x8c\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.384951 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-58cb-account-create-update-n2g6b"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.386286 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.404899 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87pgd\" (UniqueName: \"kubernetes.io/projected/d500410c-9685-4529-b187-750ea33d7b99-kube-api-access-87pgd\") pod \"barbican-db-create-s5x8c\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.475338 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd762222-8af2-4ce2-91a4-eb230aacda33-operator-scripts\") pod \"barbican-58cb-account-create-update-n2g6b\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.475475 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hfz6\" (UniqueName: \"kubernetes.io/projected/fd762222-8af2-4ce2-91a4-eb230aacda33-kube-api-access-6hfz6\") pod \"barbican-58cb-account-create-update-n2g6b\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.483890 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-sq9hc"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.485127 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.506575 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-4a1c-account-create-update-lrw4k"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.508015 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.517210 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.526543 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4a1c-account-create-update-lrw4k"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.526677 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.549968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-sq9hc"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.577704 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ed755d-307b-43d6-b84f-24c5f7284f24-operator-scripts\") pod \"cinder-4a1c-account-create-update-lrw4k\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.577834 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hfz6\" (UniqueName: \"kubernetes.io/projected/fd762222-8af2-4ce2-91a4-eb230aacda33-kube-api-access-6hfz6\") pod \"barbican-58cb-account-create-update-n2g6b\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.577984 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/224c4995-4a06-4e42-bd64-39794296a221-operator-scripts\") pod \"cinder-db-create-sq9hc\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.578028 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzrvz\" (UniqueName: \"kubernetes.io/projected/86ed755d-307b-43d6-b84f-24c5f7284f24-kube-api-access-zzrvz\") pod \"cinder-4a1c-account-create-update-lrw4k\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.578064 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd762222-8af2-4ce2-91a4-eb230aacda33-operator-scripts\") pod \"barbican-58cb-account-create-update-n2g6b\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.578145 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxhfn\" (UniqueName: \"kubernetes.io/projected/224c4995-4a06-4e42-bd64-39794296a221-kube-api-access-fxhfn\") pod \"cinder-db-create-sq9hc\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.579288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd762222-8af2-4ce2-91a4-eb230aacda33-operator-scripts\") pod \"barbican-58cb-account-create-update-n2g6b\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.631670 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hfz6\" (UniqueName: \"kubernetes.io/projected/fd762222-8af2-4ce2-91a4-eb230aacda33-kube-api-access-6hfz6\") pod \"barbican-58cb-account-create-update-n2g6b\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.647733 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-hc6p7"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.651069 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.684547 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpps8\" (UniqueName: \"kubernetes.io/projected/cf58f8ea-b763-4a18-9697-865f5741e5c7-kube-api-access-dpps8\") pod \"cloudkitty-db-create-hc6p7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.684601 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxhfn\" (UniqueName: \"kubernetes.io/projected/224c4995-4a06-4e42-bd64-39794296a221-kube-api-access-fxhfn\") pod \"cinder-db-create-sq9hc\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.684649 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ed755d-307b-43d6-b84f-24c5f7284f24-operator-scripts\") pod \"cinder-4a1c-account-create-update-lrw4k\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.684735 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/224c4995-4a06-4e42-bd64-39794296a221-operator-scripts\") pod \"cinder-db-create-sq9hc\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.684762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzrvz\" (UniqueName: \"kubernetes.io/projected/86ed755d-307b-43d6-b84f-24c5f7284f24-kube-api-access-zzrvz\") pod \"cinder-4a1c-account-create-update-lrw4k\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.684784 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf58f8ea-b763-4a18-9697-865f5741e5c7-operator-scripts\") pod \"cloudkitty-db-create-hc6p7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.686920 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/224c4995-4a06-4e42-bd64-39794296a221-operator-scripts\") pod \"cinder-db-create-sq9hc\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.687044 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ed755d-307b-43d6-b84f-24c5f7284f24-operator-scripts\") pod \"cinder-4a1c-account-create-update-lrw4k\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.690612 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.731553 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzrvz\" (UniqueName: \"kubernetes.io/projected/86ed755d-307b-43d6-b84f-24c5f7284f24-kube-api-access-zzrvz\") pod \"cinder-4a1c-account-create-update-lrw4k\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.743942 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxhfn\" (UniqueName: \"kubernetes.io/projected/224c4995-4a06-4e42-bd64-39794296a221-kube-api-access-fxhfn\") pod \"cinder-db-create-sq9hc\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.745272 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-hc6p7"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.768484 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-pk77t"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.796839 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.798331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf58f8ea-b763-4a18-9697-865f5741e5c7-operator-scripts\") pod \"cloudkitty-db-create-hc6p7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.798528 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpps8\" (UniqueName: \"kubernetes.io/projected/cf58f8ea-b763-4a18-9697-865f5741e5c7-kube-api-access-dpps8\") pod \"cloudkitty-db-create-hc6p7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.799396 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf58f8ea-b763-4a18-9697-865f5741e5c7-operator-scripts\") pod \"cloudkitty-db-create-hc6p7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.809446 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.817437 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.817690 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ph6mx" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.817839 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.818024 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.843251 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.862519 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pk77t"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.943964 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-combined-ca-bundle\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.944026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-config-data\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.944061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7znsm\" (UniqueName: \"kubernetes.io/projected/0b4489ab-f1cb-4296-b5f7-6168189d3be0-kube-api-access-7znsm\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.944166 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2ltzr"] Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.945421 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.956751 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpps8\" (UniqueName: \"kubernetes.io/projected/cf58f8ea-b763-4a18-9697-865f5741e5c7-kube-api-access-dpps8\") pod \"cloudkitty-db-create-hc6p7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:06 crc kubenswrapper[4894]: I1210 19:16:06.994449 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2ltzr"] Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.016221 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.031206 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8c45-account-create-update-6d449"] Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.032377 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.036030 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.045925 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7znsm\" (UniqueName: \"kubernetes.io/projected/0b4489ab-f1cb-4296-b5f7-6168189d3be0-kube-api-access-7znsm\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.046034 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvvqm\" (UniqueName: \"kubernetes.io/projected/ec67bfcd-9767-4e89-9683-0fe681539e23-kube-api-access-wvvqm\") pod \"neutron-db-create-2ltzr\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.046117 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-combined-ca-bundle\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.046146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-config-data\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.046170 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec67bfcd-9767-4e89-9683-0fe681539e23-operator-scripts\") pod \"neutron-db-create-2ltzr\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.052768 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-config-data\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.060839 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8c45-account-create-update-6d449"] Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.064466 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-combined-ca-bundle\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.071792 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7znsm\" (UniqueName: \"kubernetes.io/projected/0b4489ab-f1cb-4296-b5f7-6168189d3be0-kube-api-access-7znsm\") pod \"keystone-db-sync-pk77t\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.094568 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-ade3-account-create-update-cbv5v"] Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.095969 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.108048 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.120581 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-ade3-account-create-update-cbv5v"] Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.147360 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec67bfcd-9767-4e89-9683-0fe681539e23-operator-scripts\") pod \"neutron-db-create-2ltzr\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.147449 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvvqm\" (UniqueName: \"kubernetes.io/projected/ec67bfcd-9767-4e89-9683-0fe681539e23-kube-api-access-wvvqm\") pod \"neutron-db-create-2ltzr\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.147487 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znzkm\" (UniqueName: \"kubernetes.io/projected/f60159f7-610c-4277-9f46-7551f1ec247d-kube-api-access-znzkm\") pod \"neutron-8c45-account-create-update-6d449\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.147549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f60159f7-610c-4277-9f46-7551f1ec247d-operator-scripts\") pod \"neutron-8c45-account-create-update-6d449\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.148230 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec67bfcd-9767-4e89-9683-0fe681539e23-operator-scripts\") pod \"neutron-db-create-2ltzr\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.159017 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.167497 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvvqm\" (UniqueName: \"kubernetes.io/projected/ec67bfcd-9767-4e89-9683-0fe681539e23-kube-api-access-wvvqm\") pod \"neutron-db-create-2ltzr\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.249416 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f60159f7-610c-4277-9f46-7551f1ec247d-operator-scripts\") pod \"neutron-8c45-account-create-update-6d449\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.249476 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6m8s\" (UniqueName: \"kubernetes.io/projected/45f4e887-2f35-4567-b9be-931ff8e6b52c-kube-api-access-t6m8s\") pod \"cloudkitty-ade3-account-create-update-cbv5v\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.249521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45f4e887-2f35-4567-b9be-931ff8e6b52c-operator-scripts\") pod \"cloudkitty-ade3-account-create-update-cbv5v\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.249997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znzkm\" (UniqueName: \"kubernetes.io/projected/f60159f7-610c-4277-9f46-7551f1ec247d-kube-api-access-znzkm\") pod \"neutron-8c45-account-create-update-6d449\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.250198 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f60159f7-610c-4277-9f46-7551f1ec247d-operator-scripts\") pod \"neutron-8c45-account-create-update-6d449\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.267871 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znzkm\" (UniqueName: \"kubernetes.io/projected/f60159f7-610c-4277-9f46-7551f1ec247d-kube-api-access-znzkm\") pod \"neutron-8c45-account-create-update-6d449\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.310565 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.351822 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45f4e887-2f35-4567-b9be-931ff8e6b52c-operator-scripts\") pod \"cloudkitty-ade3-account-create-update-cbv5v\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.352369 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6m8s\" (UniqueName: \"kubernetes.io/projected/45f4e887-2f35-4567-b9be-931ff8e6b52c-kube-api-access-t6m8s\") pod \"cloudkitty-ade3-account-create-update-cbv5v\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.352919 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45f4e887-2f35-4567-b9be-931ff8e6b52c-operator-scripts\") pod \"cloudkitty-ade3-account-create-update-cbv5v\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.373618 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6m8s\" (UniqueName: \"kubernetes.io/projected/45f4e887-2f35-4567-b9be-931ff8e6b52c-kube-api-access-t6m8s\") pod \"cloudkitty-ade3-account-create-update-cbv5v\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.377424 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:07 crc kubenswrapper[4894]: I1210 19:16:07.447212 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.496637 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dae05eff-e7b1-460a-a374-74e698d8f0c9","Type":"ContainerStarted","Data":"d60d17feb05bdeb9fbd09d555c852b2b74e97059b16265c3543e27c63058f932"} Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.500383 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"991f2e77a693239659d867ff11515b4bd90e2452372d76afd4195802b11d0408"} Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.500410 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"ef0481734ebf13029be2846fda596ad282d3c128951d85dcffe1f2879f4d3da9"} Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.947264 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-s5x8c"] Dec 10 19:16:11 crc kubenswrapper[4894]: W1210 19:16:11.954888 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd500410c_9685_4529_b187_750ea33d7b99.slice/crio-bd763ee9e62ca7132a8cb481ec23d2f23897c7596dbd9b6715576b1d94dd363b WatchSource:0}: Error finding container bd763ee9e62ca7132a8cb481ec23d2f23897c7596dbd9b6715576b1d94dd363b: Status 404 returned error can't find the container with id bd763ee9e62ca7132a8cb481ec23d2f23897c7596dbd9b6715576b1d94dd363b Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.962347 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4a1c-account-create-update-lrw4k"] Dec 10 19:16:11 crc kubenswrapper[4894]: W1210 19:16:11.964510 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86ed755d_307b_43d6_b84f_24c5f7284f24.slice/crio-0c3f9db66676373972a976ff010ad99c5185de28c6130718604774033d56bd3d WatchSource:0}: Error finding container 0c3f9db66676373972a976ff010ad99c5185de28c6130718604774033d56bd3d: Status 404 returned error can't find the container with id 0c3f9db66676373972a976ff010ad99c5185de28c6130718604774033d56bd3d Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.974967 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8c45-account-create-update-6d449"] Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.982927 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-58cb-account-create-update-n2g6b"] Dec 10 19:16:11 crc kubenswrapper[4894]: I1210 19:16:11.990140 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pk77t"] Dec 10 19:16:11 crc kubenswrapper[4894]: W1210 19:16:11.994976 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b4489ab_f1cb_4296_b5f7_6168189d3be0.slice/crio-c20b9973a83e14b92e7f901216e9c8394fb4716401c85487df3afab0b11e9abf WatchSource:0}: Error finding container c20b9973a83e14b92e7f901216e9c8394fb4716401c85487df3afab0b11e9abf: Status 404 returned error can't find the container with id c20b9973a83e14b92e7f901216e9c8394fb4716401c85487df3afab0b11e9abf Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.163379 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-sq9hc"] Dec 10 19:16:12 crc kubenswrapper[4894]: W1210 19:16:12.168004 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45f4e887_2f35_4567_b9be_931ff8e6b52c.slice/crio-256be7dc3f6ae407dd4f83e3eeaffa1f5e82dd1944c08c6b038de8e51f054acf WatchSource:0}: Error finding container 256be7dc3f6ae407dd4f83e3eeaffa1f5e82dd1944c08c6b038de8e51f054acf: Status 404 returned error can't find the container with id 256be7dc3f6ae407dd4f83e3eeaffa1f5e82dd1944c08c6b038de8e51f054acf Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.169986 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-hc6p7"] Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.177933 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-ade3-account-create-update-cbv5v"] Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.197577 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2ltzr"] Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.511807 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2ltzr" event={"ID":"ec67bfcd-9767-4e89-9683-0fe681539e23","Type":"ContainerStarted","Data":"fa3e23c2d912272ca352df2ba1e6f91d2189f7efd5e7b51a6c364f1e58ed5b1d"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.513626 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-hc6p7" event={"ID":"cf58f8ea-b763-4a18-9697-865f5741e5c7","Type":"ContainerStarted","Data":"971eb7e85dadeaca18768502288fa10f035cfb8796decafd3a4e66520388469f"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.513666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-hc6p7" event={"ID":"cf58f8ea-b763-4a18-9697-865f5741e5c7","Type":"ContainerStarted","Data":"0809257b3027520358db9b4a12861e5ad7ad919a4621bbd373c0ca5a876bdc76"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.516199 4894 generic.go:334] "Generic (PLEG): container finished" podID="f60159f7-610c-4277-9f46-7551f1ec247d" containerID="6f705d1f9870dc89ac0338837a84a52e179f9221451404bc6d3954ce938a56d0" exitCode=0 Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.516263 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c45-account-create-update-6d449" event={"ID":"f60159f7-610c-4277-9f46-7551f1ec247d","Type":"ContainerDied","Data":"6f705d1f9870dc89ac0338837a84a52e179f9221451404bc6d3954ce938a56d0"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.516286 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c45-account-create-update-6d449" event={"ID":"f60159f7-610c-4277-9f46-7551f1ec247d","Type":"ContainerStarted","Data":"a767018840271951f2459155bb2ce35500b5626ce1023a03f7873f4ed5b92128"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.518474 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58cb-account-create-update-n2g6b" event={"ID":"fd762222-8af2-4ce2-91a4-eb230aacda33","Type":"ContainerStarted","Data":"6c6d56d2ec1fff10ae2fd378f0568fba0a7735260512f9d8fde45cfe141529a3"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.518498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58cb-account-create-update-n2g6b" event={"ID":"fd762222-8af2-4ce2-91a4-eb230aacda33","Type":"ContainerStarted","Data":"5a3d875640bdf9bcf865fad1bef4842ee7a384bfa654a386b718a6e12d7a4ab9"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.520560 4894 generic.go:334] "Generic (PLEG): container finished" podID="86ed755d-307b-43d6-b84f-24c5f7284f24" containerID="0273188ce1e0f0d6ebac343ae54ee9b66a251566c4820339e1c5badc6f6e4664" exitCode=0 Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.520584 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a1c-account-create-update-lrw4k" event={"ID":"86ed755d-307b-43d6-b84f-24c5f7284f24","Type":"ContainerDied","Data":"0273188ce1e0f0d6ebac343ae54ee9b66a251566c4820339e1c5badc6f6e4664"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.520613 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a1c-account-create-update-lrw4k" event={"ID":"86ed755d-307b-43d6-b84f-24c5f7284f24","Type":"ContainerStarted","Data":"0c3f9db66676373972a976ff010ad99c5185de28c6130718604774033d56bd3d"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.521791 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sq9hc" event={"ID":"224c4995-4a06-4e42-bd64-39794296a221","Type":"ContainerStarted","Data":"334302b0018b1479dfc8b608c7eabce46bb2181070410ee8ac65e99a53e4497e"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.536688 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"dcb75bc24d9201a4ed975ed9fb3dcbb1c3cccdfb4214631e627a7397defca56b"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.536743 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"bb1613132243d2922c073ab687545c495259d555dff85e49c56474f03df0e536"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.538547 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kxf26" event={"ID":"8f1bbcc0-1d98-452d-9b61-9caf78395cb4","Type":"ContainerStarted","Data":"c47cce434546c49a726e3eb930414b2f4c60861ec018316d9ce56d6b7190793a"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.538787 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-create-hc6p7" podStartSLOduration=6.538769251 podStartE2EDuration="6.538769251s" podCreationTimestamp="2025-12-10 19:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:12.52858483 +0000 UTC m=+1255.323432622" watchObservedRunningTime="2025-12-10 19:16:12.538769251 +0000 UTC m=+1255.333617053" Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.540602 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" event={"ID":"45f4e887-2f35-4567-b9be-931ff8e6b52c","Type":"ContainerStarted","Data":"256be7dc3f6ae407dd4f83e3eeaffa1f5e82dd1944c08c6b038de8e51f054acf"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.547336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pk77t" event={"ID":"0b4489ab-f1cb-4296-b5f7-6168189d3be0","Type":"ContainerStarted","Data":"c20b9973a83e14b92e7f901216e9c8394fb4716401c85487df3afab0b11e9abf"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.552710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s5x8c" event={"ID":"d500410c-9685-4529-b187-750ea33d7b99","Type":"ContainerStarted","Data":"6037d10d68cd85c0d7ba2af1c208575a5d2c5df33c6bbeccd2fb1e43cec12cf5"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.552749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s5x8c" event={"ID":"d500410c-9685-4529-b187-750ea33d7b99","Type":"ContainerStarted","Data":"bd763ee9e62ca7132a8cb481ec23d2f23897c7596dbd9b6715576b1d94dd363b"} Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.578009 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-58cb-account-create-update-n2g6b" podStartSLOduration=6.577980037 podStartE2EDuration="6.577980037s" podCreationTimestamp="2025-12-10 19:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:12.557281086 +0000 UTC m=+1255.352128878" watchObservedRunningTime="2025-12-10 19:16:12.577980037 +0000 UTC m=+1255.372827839" Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.595170 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" podStartSLOduration=6.595146056 podStartE2EDuration="6.595146056s" podCreationTimestamp="2025-12-10 19:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:12.58332354 +0000 UTC m=+1255.378171352" watchObservedRunningTime="2025-12-10 19:16:12.595146056 +0000 UTC m=+1255.389993848" Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.637361 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-kxf26" podStartSLOduration=3.83652743 podStartE2EDuration="18.637345291s" podCreationTimestamp="2025-12-10 19:15:54 +0000 UTC" firstStartedPulling="2025-12-10 19:15:55.809949831 +0000 UTC m=+1238.604797623" lastFinishedPulling="2025-12-10 19:16:10.610767692 +0000 UTC m=+1253.405615484" observedRunningTime="2025-12-10 19:16:12.60506506 +0000 UTC m=+1255.399912862" watchObservedRunningTime="2025-12-10 19:16:12.637345291 +0000 UTC m=+1255.432193083" Dec 10 19:16:12 crc kubenswrapper[4894]: I1210 19:16:12.647526 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-s5x8c" podStartSLOduration=6.647510402 podStartE2EDuration="6.647510402s" podCreationTimestamp="2025-12-10 19:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:12.617344628 +0000 UTC m=+1255.412192450" watchObservedRunningTime="2025-12-10 19:16:12.647510402 +0000 UTC m=+1255.442358184" Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.568567 4894 generic.go:334] "Generic (PLEG): container finished" podID="224c4995-4a06-4e42-bd64-39794296a221" containerID="620148e89fa9a80437e93e53da7fa560759cccbed1c2d01484633e1ace3ba2af" exitCode=0 Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.568781 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sq9hc" event={"ID":"224c4995-4a06-4e42-bd64-39794296a221","Type":"ContainerDied","Data":"620148e89fa9a80437e93e53da7fa560759cccbed1c2d01484633e1ace3ba2af"} Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.577295 4894 generic.go:334] "Generic (PLEG): container finished" podID="45f4e887-2f35-4567-b9be-931ff8e6b52c" containerID="7532b830b73e3d7e460d3a094e3729ab502b06106dcb41cf3785962718d445f5" exitCode=0 Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.577477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" event={"ID":"45f4e887-2f35-4567-b9be-931ff8e6b52c","Type":"ContainerDied","Data":"7532b830b73e3d7e460d3a094e3729ab502b06106dcb41cf3785962718d445f5"} Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.581546 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2ltzr" event={"ID":"ec67bfcd-9767-4e89-9683-0fe681539e23","Type":"ContainerStarted","Data":"0acc7bbf26de8adf07f305399736ee77810e3f99a01d6faf2e0bc7426c5c2841"} Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.590226 4894 generic.go:334] "Generic (PLEG): container finished" podID="fd762222-8af2-4ce2-91a4-eb230aacda33" containerID="6c6d56d2ec1fff10ae2fd378f0568fba0a7735260512f9d8fde45cfe141529a3" exitCode=0 Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.590341 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58cb-account-create-update-n2g6b" event={"ID":"fd762222-8af2-4ce2-91a4-eb230aacda33","Type":"ContainerDied","Data":"6c6d56d2ec1fff10ae2fd378f0568fba0a7735260512f9d8fde45cfe141529a3"} Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.594765 4894 generic.go:334] "Generic (PLEG): container finished" podID="cf58f8ea-b763-4a18-9697-865f5741e5c7" containerID="971eb7e85dadeaca18768502288fa10f035cfb8796decafd3a4e66520388469f" exitCode=0 Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.594860 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-hc6p7" event={"ID":"cf58f8ea-b763-4a18-9697-865f5741e5c7","Type":"ContainerDied","Data":"971eb7e85dadeaca18768502288fa10f035cfb8796decafd3a4e66520388469f"} Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.597806 4894 generic.go:334] "Generic (PLEG): container finished" podID="d500410c-9685-4529-b187-750ea33d7b99" containerID="6037d10d68cd85c0d7ba2af1c208575a5d2c5df33c6bbeccd2fb1e43cec12cf5" exitCode=0 Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.597898 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s5x8c" event={"ID":"d500410c-9685-4529-b187-750ea33d7b99","Type":"ContainerDied","Data":"6037d10d68cd85c0d7ba2af1c208575a5d2c5df33c6bbeccd2fb1e43cec12cf5"} Dec 10 19:16:13 crc kubenswrapper[4894]: I1210 19:16:13.608512 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-2ltzr" podStartSLOduration=7.608496376 podStartE2EDuration="7.608496376s" podCreationTimestamp="2025-12-10 19:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:13.605498406 +0000 UTC m=+1256.400346198" watchObservedRunningTime="2025-12-10 19:16:13.608496376 +0000 UTC m=+1256.403344168" Dec 10 19:16:14 crc kubenswrapper[4894]: I1210 19:16:14.608868 4894 generic.go:334] "Generic (PLEG): container finished" podID="ec67bfcd-9767-4e89-9683-0fe681539e23" containerID="0acc7bbf26de8adf07f305399736ee77810e3f99a01d6faf2e0bc7426c5c2841" exitCode=0 Dec 10 19:16:14 crc kubenswrapper[4894]: I1210 19:16:14.609273 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2ltzr" event={"ID":"ec67bfcd-9767-4e89-9683-0fe681539e23","Type":"ContainerDied","Data":"0acc7bbf26de8adf07f305399736ee77810e3f99a01d6faf2e0bc7426c5c2841"} Dec 10 19:16:14 crc kubenswrapper[4894]: I1210 19:16:14.619769 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"0c5a3c4b0b6e71bddd8343cc7747e0d23a380d26522434e78cdfdf87f8e00389"} Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.643127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-58cb-account-create-update-n2g6b" event={"ID":"fd762222-8af2-4ce2-91a4-eb230aacda33","Type":"ContainerDied","Data":"5a3d875640bdf9bcf865fad1bef4842ee7a384bfa654a386b718a6e12d7a4ab9"} Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.643738 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a3d875640bdf9bcf865fad1bef4842ee7a384bfa654a386b718a6e12d7a4ab9" Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.889313 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.943925 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd762222-8af2-4ce2-91a4-eb230aacda33-operator-scripts\") pod \"fd762222-8af2-4ce2-91a4-eb230aacda33\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.944226 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hfz6\" (UniqueName: \"kubernetes.io/projected/fd762222-8af2-4ce2-91a4-eb230aacda33-kube-api-access-6hfz6\") pod \"fd762222-8af2-4ce2-91a4-eb230aacda33\" (UID: \"fd762222-8af2-4ce2-91a4-eb230aacda33\") " Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.946426 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd762222-8af2-4ce2-91a4-eb230aacda33-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd762222-8af2-4ce2-91a4-eb230aacda33" (UID: "fd762222-8af2-4ce2-91a4-eb230aacda33"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.946713 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:16 crc kubenswrapper[4894]: I1210 19:16:16.949393 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd762222-8af2-4ce2-91a4-eb230aacda33-kube-api-access-6hfz6" (OuterVolumeSpecName: "kube-api-access-6hfz6") pod "fd762222-8af2-4ce2-91a4-eb230aacda33" (UID: "fd762222-8af2-4ce2-91a4-eb230aacda33"). InnerVolumeSpecName "kube-api-access-6hfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.025317 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.033517 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.048163 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/224c4995-4a06-4e42-bd64-39794296a221-operator-scripts\") pod \"224c4995-4a06-4e42-bd64-39794296a221\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.048207 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxhfn\" (UniqueName: \"kubernetes.io/projected/224c4995-4a06-4e42-bd64-39794296a221-kube-api-access-fxhfn\") pod \"224c4995-4a06-4e42-bd64-39794296a221\" (UID: \"224c4995-4a06-4e42-bd64-39794296a221\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.048621 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hfz6\" (UniqueName: \"kubernetes.io/projected/fd762222-8af2-4ce2-91a4-eb230aacda33-kube-api-access-6hfz6\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.048634 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd762222-8af2-4ce2-91a4-eb230aacda33-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.053338 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224c4995-4a06-4e42-bd64-39794296a221-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "224c4995-4a06-4e42-bd64-39794296a221" (UID: "224c4995-4a06-4e42-bd64-39794296a221"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.086129 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224c4995-4a06-4e42-bd64-39794296a221-kube-api-access-fxhfn" (OuterVolumeSpecName: "kube-api-access-fxhfn") pod "224c4995-4a06-4e42-bd64-39794296a221" (UID: "224c4995-4a06-4e42-bd64-39794296a221"). InnerVolumeSpecName "kube-api-access-fxhfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.096623 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.132105 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.149695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45f4e887-2f35-4567-b9be-931ff8e6b52c-operator-scripts\") pod \"45f4e887-2f35-4567-b9be-931ff8e6b52c\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.149739 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpps8\" (UniqueName: \"kubernetes.io/projected/cf58f8ea-b763-4a18-9697-865f5741e5c7-kube-api-access-dpps8\") pod \"cf58f8ea-b763-4a18-9697-865f5741e5c7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.149803 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6m8s\" (UniqueName: \"kubernetes.io/projected/45f4e887-2f35-4567-b9be-931ff8e6b52c-kube-api-access-t6m8s\") pod \"45f4e887-2f35-4567-b9be-931ff8e6b52c\" (UID: \"45f4e887-2f35-4567-b9be-931ff8e6b52c\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.149896 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ed755d-307b-43d6-b84f-24c5f7284f24-operator-scripts\") pod \"86ed755d-307b-43d6-b84f-24c5f7284f24\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.149948 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf58f8ea-b763-4a18-9697-865f5741e5c7-operator-scripts\") pod \"cf58f8ea-b763-4a18-9697-865f5741e5c7\" (UID: \"cf58f8ea-b763-4a18-9697-865f5741e5c7\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.149996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzrvz\" (UniqueName: \"kubernetes.io/projected/86ed755d-307b-43d6-b84f-24c5f7284f24-kube-api-access-zzrvz\") pod \"86ed755d-307b-43d6-b84f-24c5f7284f24\" (UID: \"86ed755d-307b-43d6-b84f-24c5f7284f24\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.150411 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/224c4995-4a06-4e42-bd64-39794296a221-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.150429 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxhfn\" (UniqueName: \"kubernetes.io/projected/224c4995-4a06-4e42-bd64-39794296a221-kube-api-access-fxhfn\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.151004 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86ed755d-307b-43d6-b84f-24c5f7284f24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86ed755d-307b-43d6-b84f-24c5f7284f24" (UID: "86ed755d-307b-43d6-b84f-24c5f7284f24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.153688 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf58f8ea-b763-4a18-9697-865f5741e5c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf58f8ea-b763-4a18-9697-865f5741e5c7" (UID: "cf58f8ea-b763-4a18-9697-865f5741e5c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.154491 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f4e887-2f35-4567-b9be-931ff8e6b52c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45f4e887-2f35-4567-b9be-931ff8e6b52c" (UID: "45f4e887-2f35-4567-b9be-931ff8e6b52c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.156881 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f4e887-2f35-4567-b9be-931ff8e6b52c-kube-api-access-t6m8s" (OuterVolumeSpecName: "kube-api-access-t6m8s") pod "45f4e887-2f35-4567-b9be-931ff8e6b52c" (UID: "45f4e887-2f35-4567-b9be-931ff8e6b52c"). InnerVolumeSpecName "kube-api-access-t6m8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.165490 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.167641 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ed755d-307b-43d6-b84f-24c5f7284f24-kube-api-access-zzrvz" (OuterVolumeSpecName: "kube-api-access-zzrvz") pod "86ed755d-307b-43d6-b84f-24c5f7284f24" (UID: "86ed755d-307b-43d6-b84f-24c5f7284f24"). InnerVolumeSpecName "kube-api-access-zzrvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.168770 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.168995 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf58f8ea-b763-4a18-9697-865f5741e5c7-kube-api-access-dpps8" (OuterVolumeSpecName: "kube-api-access-dpps8") pod "cf58f8ea-b763-4a18-9697-865f5741e5c7" (UID: "cf58f8ea-b763-4a18-9697-865f5741e5c7"). InnerVolumeSpecName "kube-api-access-dpps8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.251199 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87pgd\" (UniqueName: \"kubernetes.io/projected/d500410c-9685-4529-b187-750ea33d7b99-kube-api-access-87pgd\") pod \"d500410c-9685-4529-b187-750ea33d7b99\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.251279 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znzkm\" (UniqueName: \"kubernetes.io/projected/f60159f7-610c-4277-9f46-7551f1ec247d-kube-api-access-znzkm\") pod \"f60159f7-610c-4277-9f46-7551f1ec247d\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.251322 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec67bfcd-9767-4e89-9683-0fe681539e23-operator-scripts\") pod \"ec67bfcd-9767-4e89-9683-0fe681539e23\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.251353 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d500410c-9685-4529-b187-750ea33d7b99-operator-scripts\") pod \"d500410c-9685-4529-b187-750ea33d7b99\" (UID: \"d500410c-9685-4529-b187-750ea33d7b99\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.251458 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvvqm\" (UniqueName: \"kubernetes.io/projected/ec67bfcd-9767-4e89-9683-0fe681539e23-kube-api-access-wvvqm\") pod \"ec67bfcd-9767-4e89-9683-0fe681539e23\" (UID: \"ec67bfcd-9767-4e89-9683-0fe681539e23\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.251493 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f60159f7-610c-4277-9f46-7551f1ec247d-operator-scripts\") pod \"f60159f7-610c-4277-9f46-7551f1ec247d\" (UID: \"f60159f7-610c-4277-9f46-7551f1ec247d\") " Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252030 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d500410c-9685-4529-b187-750ea33d7b99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d500410c-9685-4529-b187-750ea33d7b99" (UID: "d500410c-9685-4529-b187-750ea33d7b99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252239 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d500410c-9685-4529-b187-750ea33d7b99-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252271 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzrvz\" (UniqueName: \"kubernetes.io/projected/86ed755d-307b-43d6-b84f-24c5f7284f24-kube-api-access-zzrvz\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252288 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45f4e887-2f35-4567-b9be-931ff8e6b52c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252333 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpps8\" (UniqueName: \"kubernetes.io/projected/cf58f8ea-b763-4a18-9697-865f5741e5c7-kube-api-access-dpps8\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252347 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6m8s\" (UniqueName: \"kubernetes.io/projected/45f4e887-2f35-4567-b9be-931ff8e6b52c-kube-api-access-t6m8s\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252359 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ed755d-307b-43d6-b84f-24c5f7284f24-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252370 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf58f8ea-b763-4a18-9697-865f5741e5c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.252986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec67bfcd-9767-4e89-9683-0fe681539e23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec67bfcd-9767-4e89-9683-0fe681539e23" (UID: "ec67bfcd-9767-4e89-9683-0fe681539e23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.253912 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60159f7-610c-4277-9f46-7551f1ec247d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f60159f7-610c-4277-9f46-7551f1ec247d" (UID: "f60159f7-610c-4277-9f46-7551f1ec247d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.254128 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60159f7-610c-4277-9f46-7551f1ec247d-kube-api-access-znzkm" (OuterVolumeSpecName: "kube-api-access-znzkm") pod "f60159f7-610c-4277-9f46-7551f1ec247d" (UID: "f60159f7-610c-4277-9f46-7551f1ec247d"). InnerVolumeSpecName "kube-api-access-znzkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.254936 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec67bfcd-9767-4e89-9683-0fe681539e23-kube-api-access-wvvqm" (OuterVolumeSpecName: "kube-api-access-wvvqm") pod "ec67bfcd-9767-4e89-9683-0fe681539e23" (UID: "ec67bfcd-9767-4e89-9683-0fe681539e23"). InnerVolumeSpecName "kube-api-access-wvvqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.263724 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d500410c-9685-4529-b187-750ea33d7b99-kube-api-access-87pgd" (OuterVolumeSpecName: "kube-api-access-87pgd") pod "d500410c-9685-4529-b187-750ea33d7b99" (UID: "d500410c-9685-4529-b187-750ea33d7b99"). InnerVolumeSpecName "kube-api-access-87pgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.354609 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvvqm\" (UniqueName: \"kubernetes.io/projected/ec67bfcd-9767-4e89-9683-0fe681539e23-kube-api-access-wvvqm\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.354648 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f60159f7-610c-4277-9f46-7551f1ec247d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.354662 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87pgd\" (UniqueName: \"kubernetes.io/projected/d500410c-9685-4529-b187-750ea33d7b99-kube-api-access-87pgd\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.354676 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znzkm\" (UniqueName: \"kubernetes.io/projected/f60159f7-610c-4277-9f46-7551f1ec247d-kube-api-access-znzkm\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.354694 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec67bfcd-9767-4e89-9683-0fe681539e23-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.665480 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2ltzr" event={"ID":"ec67bfcd-9767-4e89-9683-0fe681539e23","Type":"ContainerDied","Data":"fa3e23c2d912272ca352df2ba1e6f91d2189f7efd5e7b51a6c364f1e58ed5b1d"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.666612 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa3e23c2d912272ca352df2ba1e6f91d2189f7efd5e7b51a6c364f1e58ed5b1d" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.665959 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2ltzr" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.674463 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-hc6p7" event={"ID":"cf58f8ea-b763-4a18-9697-865f5741e5c7","Type":"ContainerDied","Data":"0809257b3027520358db9b4a12861e5ad7ad919a4621bbd373c0ca5a876bdc76"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.674498 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0809257b3027520358db9b4a12861e5ad7ad919a4621bbd373c0ca5a876bdc76" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.674567 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-hc6p7" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.681075 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s5x8c" event={"ID":"d500410c-9685-4529-b187-750ea33d7b99","Type":"ContainerDied","Data":"bd763ee9e62ca7132a8cb481ec23d2f23897c7596dbd9b6715576b1d94dd363b"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.681106 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s5x8c" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.681118 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd763ee9e62ca7132a8cb481ec23d2f23897c7596dbd9b6715576b1d94dd363b" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.682972 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sq9hc" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.683141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sq9hc" event={"ID":"224c4995-4a06-4e42-bd64-39794296a221","Type":"ContainerDied","Data":"334302b0018b1479dfc8b608c7eabce46bb2181070410ee8ac65e99a53e4497e"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.683176 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="334302b0018b1479dfc8b608c7eabce46bb2181070410ee8ac65e99a53e4497e" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.685352 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dae05eff-e7b1-460a-a374-74e698d8f0c9","Type":"ContainerStarted","Data":"b731148f44a1e6fb0464e031636be456861ec2fda518ec2040a9de9d2bb5e837"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.685381 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dae05eff-e7b1-460a-a374-74e698d8f0c9","Type":"ContainerStarted","Data":"673fa8d4f862b5d8104672261f322454ea2da86ca4c362115d2d141738013791"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.699915 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"f7111abe5eff9063226ea89beddd1893c262efa7dc6103811865014932e5353d"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.699973 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"8ac65f7eb343666a8ac6240370ccc893aac64f82e0313019c9d5d12c0d3230fd"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.699983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"242e14b83712e3e6eb79bc9d716ad89940128026cff008d4ff54385ce1f27d57"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.699991 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"15909205ca9875cfd7543701ad51c25a4942209ecb6767dcd58b17cb74881ffd"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.703401 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c45-account-create-update-6d449" event={"ID":"f60159f7-610c-4277-9f46-7551f1ec247d","Type":"ContainerDied","Data":"a767018840271951f2459155bb2ce35500b5626ce1023a03f7873f4ed5b92128"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.703425 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a767018840271951f2459155bb2ce35500b5626ce1023a03f7873f4ed5b92128" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.703489 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c45-account-create-update-6d449" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.716236 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=23.716212299 podStartE2EDuration="23.716212299s" podCreationTimestamp="2025-12-10 19:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:17.71135795 +0000 UTC m=+1260.506205762" watchObservedRunningTime="2025-12-10 19:16:17.716212299 +0000 UTC m=+1260.511060101" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.716987 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4a1c-account-create-update-lrw4k" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.717217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4a1c-account-create-update-lrw4k" event={"ID":"86ed755d-307b-43d6-b84f-24c5f7284f24","Type":"ContainerDied","Data":"0c3f9db66676373972a976ff010ad99c5185de28c6130718604774033d56bd3d"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.717372 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c3f9db66676373972a976ff010ad99c5185de28c6130718604774033d56bd3d" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.720361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" event={"ID":"45f4e887-2f35-4567-b9be-931ff8e6b52c","Type":"ContainerDied","Data":"256be7dc3f6ae407dd4f83e3eeaffa1f5e82dd1944c08c6b038de8e51f054acf"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.720399 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="256be7dc3f6ae407dd4f83e3eeaffa1f5e82dd1944c08c6b038de8e51f054acf" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.720454 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-ade3-account-create-update-cbv5v" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.725770 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-58cb-account-create-update-n2g6b" Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.725743 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pk77t" event={"ID":"0b4489ab-f1cb-4296-b5f7-6168189d3be0","Type":"ContainerStarted","Data":"1b334d8718fa9f84010e45a2ba7dbb672f710af8d56ea574e4b19198bd4f08a5"} Dec 10 19:16:17 crc kubenswrapper[4894]: I1210 19:16:17.748579 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-pk77t" podStartSLOduration=7.046238327 podStartE2EDuration="11.748562042s" podCreationTimestamp="2025-12-10 19:16:06 +0000 UTC" firstStartedPulling="2025-12-10 19:16:12.003641327 +0000 UTC m=+1254.798489119" lastFinishedPulling="2025-12-10 19:16:16.705965042 +0000 UTC m=+1259.500812834" observedRunningTime="2025-12-10 19:16:17.743887968 +0000 UTC m=+1260.538735780" watchObservedRunningTime="2025-12-10 19:16:17.748562042 +0000 UTC m=+1260.543409834" Dec 10 19:16:18 crc kubenswrapper[4894]: I1210 19:16:18.746665 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"f93c17ed57991c6d0278f8f48f46a1fe49b80e0e07e8c5f92d2bf98776595c02"} Dec 10 19:16:18 crc kubenswrapper[4894]: I1210 19:16:18.747157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5eac36f4-3cbc-48c1-9c6a-4e79de593d04","Type":"ContainerStarted","Data":"be7c2d1acf0d24c31d86cc3aa371ebcb6b0a8d7d1f5880211e273e4ed49007af"} Dec 10 19:16:18 crc kubenswrapper[4894]: I1210 19:16:18.799746 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.643336988 podStartE2EDuration="57.79971372s" podCreationTimestamp="2025-12-10 19:15:21 +0000 UTC" firstStartedPulling="2025-12-10 19:15:55.331884128 +0000 UTC m=+1238.126731920" lastFinishedPulling="2025-12-10 19:16:13.48826086 +0000 UTC m=+1256.283108652" observedRunningTime="2025-12-10 19:16:18.789138659 +0000 UTC m=+1261.583986521" watchObservedRunningTime="2025-12-10 19:16:18.79971372 +0000 UTC m=+1261.594561592" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.076806 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-vz5cc"] Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.077519 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d500410c-9685-4529-b187-750ea33d7b99" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.077607 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d500410c-9685-4529-b187-750ea33d7b99" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.077688 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224c4995-4a06-4e42-bd64-39794296a221" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.077758 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="224c4995-4a06-4e42-bd64-39794296a221" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.077825 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f4e887-2f35-4567-b9be-931ff8e6b52c" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.077914 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f4e887-2f35-4567-b9be-931ff8e6b52c" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.077998 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec67bfcd-9767-4e89-9683-0fe681539e23" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.078057 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec67bfcd-9767-4e89-9683-0fe681539e23" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.078127 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd762222-8af2-4ce2-91a4-eb230aacda33" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.078201 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd762222-8af2-4ce2-91a4-eb230aacda33" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.078269 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60159f7-610c-4277-9f46-7551f1ec247d" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.078332 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60159f7-610c-4277-9f46-7551f1ec247d" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.078408 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf58f8ea-b763-4a18-9697-865f5741e5c7" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.078486 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf58f8ea-b763-4a18-9697-865f5741e5c7" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: E1210 19:16:19.078552 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ed755d-307b-43d6-b84f-24c5f7284f24" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.078622 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ed755d-307b-43d6-b84f-24c5f7284f24" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.078950 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ed755d-307b-43d6-b84f-24c5f7284f24" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079051 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd762222-8af2-4ce2-91a4-eb230aacda33" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079119 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="45f4e887-2f35-4567-b9be-931ff8e6b52c" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079177 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="224c4995-4a06-4e42-bd64-39794296a221" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079230 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d500410c-9685-4529-b187-750ea33d7b99" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079313 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec67bfcd-9767-4e89-9683-0fe681539e23" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079383 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf58f8ea-b763-4a18-9697-865f5741e5c7" containerName="mariadb-database-create" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.079443 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f60159f7-610c-4277-9f46-7551f1ec247d" containerName="mariadb-account-create-update" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.080656 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.082913 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.100326 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-vz5cc"] Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.195144 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-config\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.195736 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.195890 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.196419 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpzxr\" (UniqueName: \"kubernetes.io/projected/b0348d25-e268-4a8d-a805-c3a33ce28b39-kube-api-access-wpzxr\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.196716 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.196847 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.298647 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.298709 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.298744 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-config\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.298762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.298791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.298917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpzxr\" (UniqueName: \"kubernetes.io/projected/b0348d25-e268-4a8d-a805-c3a33ce28b39-kube-api-access-wpzxr\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.299710 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.299833 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.299878 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-config\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.300152 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.300507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.318278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpzxr\" (UniqueName: \"kubernetes.io/projected/b0348d25-e268-4a8d-a805-c3a33ce28b39-kube-api-access-wpzxr\") pod \"dnsmasq-dns-6d5b6d6b67-vz5cc\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.400843 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.901323 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-vz5cc"] Dec 10 19:16:19 crc kubenswrapper[4894]: W1210 19:16:19.907582 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0348d25_e268_4a8d_a805_c3a33ce28b39.slice/crio-2e9d965dc895a952119ac3bb39cbd4092726b1d652db0e27bb07f38cab82caaa WatchSource:0}: Error finding container 2e9d965dc895a952119ac3bb39cbd4092726b1d652db0e27bb07f38cab82caaa: Status 404 returned error can't find the container with id 2e9d965dc895a952119ac3bb39cbd4092726b1d652db0e27bb07f38cab82caaa Dec 10 19:16:19 crc kubenswrapper[4894]: I1210 19:16:19.961540 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.766520 4894 generic.go:334] "Generic (PLEG): container finished" podID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerID="efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0" exitCode=0 Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.766629 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" event={"ID":"b0348d25-e268-4a8d-a805-c3a33ce28b39","Type":"ContainerDied","Data":"efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0"} Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.766676 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" event={"ID":"b0348d25-e268-4a8d-a805-c3a33ce28b39","Type":"ContainerStarted","Data":"2e9d965dc895a952119ac3bb39cbd4092726b1d652db0e27bb07f38cab82caaa"} Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.772285 4894 generic.go:334] "Generic (PLEG): container finished" podID="8f1bbcc0-1d98-452d-9b61-9caf78395cb4" containerID="c47cce434546c49a726e3eb930414b2f4c60861ec018316d9ce56d6b7190793a" exitCode=0 Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.772454 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kxf26" event={"ID":"8f1bbcc0-1d98-452d-9b61-9caf78395cb4","Type":"ContainerDied","Data":"c47cce434546c49a726e3eb930414b2f4c60861ec018316d9ce56d6b7190793a"} Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.785325 4894 generic.go:334] "Generic (PLEG): container finished" podID="0b4489ab-f1cb-4296-b5f7-6168189d3be0" containerID="1b334d8718fa9f84010e45a2ba7dbb672f710af8d56ea574e4b19198bd4f08a5" exitCode=0 Dec 10 19:16:20 crc kubenswrapper[4894]: I1210 19:16:20.785391 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pk77t" event={"ID":"0b4489ab-f1cb-4296-b5f7-6168189d3be0","Type":"ContainerDied","Data":"1b334d8718fa9f84010e45a2ba7dbb672f710af8d56ea574e4b19198bd4f08a5"} Dec 10 19:16:21 crc kubenswrapper[4894]: I1210 19:16:21.794263 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" event={"ID":"b0348d25-e268-4a8d-a805-c3a33ce28b39","Type":"ContainerStarted","Data":"6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32"} Dec 10 19:16:21 crc kubenswrapper[4894]: I1210 19:16:21.816259 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" podStartSLOduration=2.816235567 podStartE2EDuration="2.816235567s" podCreationTimestamp="2025-12-10 19:16:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:21.81525857 +0000 UTC m=+1264.610106382" watchObservedRunningTime="2025-12-10 19:16:21.816235567 +0000 UTC m=+1264.611083359" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.172165 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.356224 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-combined-ca-bundle\") pod \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.356447 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-config-data\") pod \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.356567 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7znsm\" (UniqueName: \"kubernetes.io/projected/0b4489ab-f1cb-4296-b5f7-6168189d3be0-kube-api-access-7znsm\") pod \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\" (UID: \"0b4489ab-f1cb-4296-b5f7-6168189d3be0\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.372095 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b4489ab-f1cb-4296-b5f7-6168189d3be0-kube-api-access-7znsm" (OuterVolumeSpecName: "kube-api-access-7znsm") pod "0b4489ab-f1cb-4296-b5f7-6168189d3be0" (UID: "0b4489ab-f1cb-4296-b5f7-6168189d3be0"). InnerVolumeSpecName "kube-api-access-7znsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.398696 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b4489ab-f1cb-4296-b5f7-6168189d3be0" (UID: "0b4489ab-f1cb-4296-b5f7-6168189d3be0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.418071 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-config-data" (OuterVolumeSpecName: "config-data") pod "0b4489ab-f1cb-4296-b5f7-6168189d3be0" (UID: "0b4489ab-f1cb-4296-b5f7-6168189d3be0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.458421 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7znsm\" (UniqueName: \"kubernetes.io/projected/0b4489ab-f1cb-4296-b5f7-6168189d3be0-kube-api-access-7znsm\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.458449 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.458459 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4489ab-f1cb-4296-b5f7-6168189d3be0-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.527441 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kxf26" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.663690 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-config-data\") pod \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.664186 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbk2r\" (UniqueName: \"kubernetes.io/projected/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-kube-api-access-rbk2r\") pod \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.664276 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-db-sync-config-data\") pod \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.664331 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-combined-ca-bundle\") pod \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\" (UID: \"8f1bbcc0-1d98-452d-9b61-9caf78395cb4\") " Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.668467 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-kube-api-access-rbk2r" (OuterVolumeSpecName: "kube-api-access-rbk2r") pod "8f1bbcc0-1d98-452d-9b61-9caf78395cb4" (UID: "8f1bbcc0-1d98-452d-9b61-9caf78395cb4"). InnerVolumeSpecName "kube-api-access-rbk2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.670080 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8f1bbcc0-1d98-452d-9b61-9caf78395cb4" (UID: "8f1bbcc0-1d98-452d-9b61-9caf78395cb4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.706191 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f1bbcc0-1d98-452d-9b61-9caf78395cb4" (UID: "8f1bbcc0-1d98-452d-9b61-9caf78395cb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.720840 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-config-data" (OuterVolumeSpecName: "config-data") pod "8f1bbcc0-1d98-452d-9b61-9caf78395cb4" (UID: "8f1bbcc0-1d98-452d-9b61-9caf78395cb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.767162 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbk2r\" (UniqueName: \"kubernetes.io/projected/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-kube-api-access-rbk2r\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.767417 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.767426 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.767435 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f1bbcc0-1d98-452d-9b61-9caf78395cb4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.828318 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kxf26" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.828294 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kxf26" event={"ID":"8f1bbcc0-1d98-452d-9b61-9caf78395cb4","Type":"ContainerDied","Data":"d672c0a49a4f0c2650f1195399654d8af3e32a4cb5f2d8fed97321d6f3c79880"} Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.828586 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d672c0a49a4f0c2650f1195399654d8af3e32a4cb5f2d8fed97321d6f3c79880" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.831290 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pk77t" event={"ID":"0b4489ab-f1cb-4296-b5f7-6168189d3be0","Type":"ContainerDied","Data":"c20b9973a83e14b92e7f901216e9c8394fb4716401c85487df3afab0b11e9abf"} Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.831345 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c20b9973a83e14b92e7f901216e9c8394fb4716401c85487df3afab0b11e9abf" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.831495 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pk77t" Dec 10 19:16:22 crc kubenswrapper[4894]: I1210 19:16:22.831878 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.161477 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8w9m9"] Dec 10 19:16:23 crc kubenswrapper[4894]: E1210 19:16:23.161863 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f1bbcc0-1d98-452d-9b61-9caf78395cb4" containerName="glance-db-sync" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.161879 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f1bbcc0-1d98-452d-9b61-9caf78395cb4" containerName="glance-db-sync" Dec 10 19:16:23 crc kubenswrapper[4894]: E1210 19:16:23.161899 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b4489ab-f1cb-4296-b5f7-6168189d3be0" containerName="keystone-db-sync" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.161906 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b4489ab-f1cb-4296-b5f7-6168189d3be0" containerName="keystone-db-sync" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.162071 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b4489ab-f1cb-4296-b5f7-6168189d3be0" containerName="keystone-db-sync" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.162093 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f1bbcc0-1d98-452d-9b61-9caf78395cb4" containerName="glance-db-sync" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.162711 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.164601 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.165428 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ph6mx" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.165724 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.165975 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.166149 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.188829 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-vz5cc"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.210503 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8w9m9"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.256395 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-hvdqp"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.258329 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.270207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-hvdqp"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.276476 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-config-data\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.276535 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-scripts\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.276565 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vtm\" (UniqueName: \"kubernetes.io/projected/b1ad5751-b675-40b2-bae2-2e0de76257a3-kube-api-access-r4vtm\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.276615 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-combined-ca-bundle\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.276644 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-fernet-keys\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.276704 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-credential-keys\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.363359 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-hvdqp"] Dec 10 19:16:23 crc kubenswrapper[4894]: E1210 19:16:23.364089 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-tvt6t ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" podUID="3b56057d-1ca6-4fc5-a104-37634f5e19d3" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378280 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-config\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378320 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378367 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-config-data\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-scripts\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vtm\" (UniqueName: \"kubernetes.io/projected/b1ad5751-b675-40b2-bae2-2e0de76257a3-kube-api-access-r4vtm\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378662 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378726 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378869 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-combined-ca-bundle\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378924 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-fernet-keys\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.378999 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvt6t\" (UniqueName: \"kubernetes.io/projected/3b56057d-1ca6-4fc5-a104-37634f5e19d3-kube-api-access-tvt6t\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.379112 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-credential-keys\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.386334 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.388524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-fernet-keys\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.388727 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.391308 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-credential-keys\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.392441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-config-data\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.396294 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.396490 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.397491 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-combined-ca-bundle\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.405297 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-scripts\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.433057 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.464459 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vtm\" (UniqueName: \"kubernetes.io/projected/b1ad5751-b675-40b2-bae2-2e0de76257a3-kube-api-access-r4vtm\") pod \"keystone-bootstrap-8w9m9\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.480650 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.481551 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chfwn"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.483990 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-config\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484052 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484096 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-log-httpd\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484124 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dggnm\" (UniqueName: \"kubernetes.io/projected/fe2420b5-49dd-45eb-875e-81014391d3a0-kube-api-access-dggnm\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484157 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484182 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484201 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-scripts\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-config-data\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484253 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-run-httpd\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484307 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvt6t\" (UniqueName: \"kubernetes.io/projected/3b56057d-1ca6-4fc5-a104-37634f5e19d3-kube-api-access-tvt6t\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.484363 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.485258 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.485664 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-config\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.486410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.487286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.487432 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.526316 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chfwn"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.526895 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvt6t\" (UniqueName: \"kubernetes.io/projected/3b56057d-1ca6-4fc5-a104-37634f5e19d3-kube-api-access-tvt6t\") pod \"dnsmasq-dns-6f8c45789f-hvdqp\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.586186 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jk98n"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587301 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587335 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-log-httpd\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587434 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587454 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dggnm\" (UniqueName: \"kubernetes.io/projected/fe2420b5-49dd-45eb-875e-81014391d3a0-kube-api-access-dggnm\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587500 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-scripts\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-config-data\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrdxm\" (UniqueName: \"kubernetes.io/projected/395f1d4d-ca65-4eb5-868a-e12ce3573151-kube-api-access-lrdxm\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587567 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587572 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587581 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-run-httpd\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587600 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.587656 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-config\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.591256 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-log-httpd\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.591494 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.591711 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-72ww2" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.592178 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-run-httpd\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.599148 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-scripts\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.604873 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.604941 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-5bq9r"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.606025 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.608500 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ktrjs" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.608695 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.611028 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.614535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.616402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-config-data\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.626418 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4gmfn"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.627887 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.628719 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.646689 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.646888 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-jzvfc" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.647789 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.657529 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jk98n"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.675777 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dggnm\" (UniqueName: \"kubernetes.io/projected/fe2420b5-49dd-45eb-875e-81014391d3a0-kube-api-access-dggnm\") pod \"ceilometer-0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-config\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689202 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-config\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689289 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npns8\" (UniqueName: \"kubernetes.io/projected/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-kube-api-access-npns8\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689333 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689374 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-combined-ca-bundle\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689459 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrdxm\" (UniqueName: \"kubernetes.io/projected/395f1d4d-ca65-4eb5-868a-e12ce3573151-kube-api-access-lrdxm\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.689498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.690567 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.690645 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.691089 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-config\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.691218 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.696424 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.714088 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5bq9r"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.725018 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4gmfn"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.734526 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrdxm\" (UniqueName: \"kubernetes.io/projected/395f1d4d-ca65-4eb5-868a-e12ce3573151-kube-api-access-lrdxm\") pod \"dnsmasq-dns-6c9c9f998c-chfwn\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.781836 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5765m"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.783108 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.785461 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rt8vq" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.785839 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.799230 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-config-data\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.799312 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcdch\" (UniqueName: \"kubernetes.io/projected/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-kube-api-access-gcdch\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.799335 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-scripts\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.799379 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a781080-8482-4915-9d13-b88d1785f7f3-logs\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.799411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-config\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.800605 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-combined-ca-bundle\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.800782 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-etc-machine-id\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.800802 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-scripts\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.800880 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.800912 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-config-data\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.800962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npns8\" (UniqueName: \"kubernetes.io/projected/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-kube-api-access-npns8\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.801020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knrsg\" (UniqueName: \"kubernetes.io/projected/7a781080-8482-4915-9d13-b88d1785f7f3-kube-api-access-knrsg\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.801092 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-combined-ca-bundle\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.801177 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-db-sync-config-data\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.805404 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chfwn"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.813840 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.817224 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-config\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.820164 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-combined-ca-bundle\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.832187 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npns8\" (UniqueName: \"kubernetes.io/projected/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-kube-api-access-npns8\") pod \"neutron-db-sync-jk98n\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.857124 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.869158 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5765m"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.879683 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.894322 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.905981 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-9tcn9"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.906961 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-db-sync-config-data\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907009 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-db-sync-config-data\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907170 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-config-data\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907251 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcdch\" (UniqueName: \"kubernetes.io/projected/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-kube-api-access-gcdch\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907292 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-scripts\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907322 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a781080-8482-4915-9d13-b88d1785f7f3-logs\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907451 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-combined-ca-bundle\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907484 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-etc-machine-id\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-scripts\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907564 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-config-data\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907686 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdvk8\" (UniqueName: \"kubernetes.io/projected/bedfd09b-1fdb-43e9-88d7-47fd25103af6-kube-api-access-hdvk8\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907718 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knrsg\" (UniqueName: \"kubernetes.io/projected/7a781080-8482-4915-9d13-b88d1785f7f3-kube-api-access-knrsg\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907736 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-combined-ca-bundle\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.907831 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.908697 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-etc-machine-id\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.911965 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a781080-8482-4915-9d13-b88d1785f7f3-logs\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.915643 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-combined-ca-bundle\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.917348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.920824 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-db-sync-config-data\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.920944 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-vprjr"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.924219 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-scripts\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.924994 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-config-data\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.930546 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-config-data\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.937563 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-scripts\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.937784 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-9tcn9"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.937897 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.940800 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.940995 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-gf8xv" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.941136 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcdch\" (UniqueName: \"kubernetes.io/projected/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-kube-api-access-gcdch\") pod \"cinder-db-sync-5bq9r\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.941182 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.943144 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.943642 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.947806 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knrsg\" (UniqueName: \"kubernetes.io/projected/7a781080-8482-4915-9d13-b88d1785f7f3-kube-api-access-knrsg\") pod \"placement-db-sync-4gmfn\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.963204 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-vprjr"] Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.967792 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:16:23 crc kubenswrapper[4894]: I1210 19:16:23.991713 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4gmfn" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012116 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-nb\") pod \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvt6t\" (UniqueName: \"kubernetes.io/projected/3b56057d-1ca6-4fc5-a104-37634f5e19d3-kube-api-access-tvt6t\") pod \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012199 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-swift-storage-0\") pod \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012246 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-config\") pod \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012292 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-sb\") pod \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012354 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-svc\") pod \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\" (UID: \"3b56057d-1ca6-4fc5-a104-37634f5e19d3\") " Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012589 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012627 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012665 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-config\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012758 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzjgh\" (UniqueName: \"kubernetes.io/projected/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-kube-api-access-kzjgh\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012798 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012822 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdvk8\" (UniqueName: \"kubernetes.io/projected/bedfd09b-1fdb-43e9-88d7-47fd25103af6-kube-api-access-hdvk8\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012845 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012886 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-combined-ca-bundle\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.012931 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-db-sync-config-data\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.013282 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3b56057d-1ca6-4fc5-a104-37634f5e19d3" (UID: "3b56057d-1ca6-4fc5-a104-37634f5e19d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.014459 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b56057d-1ca6-4fc5-a104-37634f5e19d3" (UID: "3b56057d-1ca6-4fc5-a104-37634f5e19d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.014889 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3b56057d-1ca6-4fc5-a104-37634f5e19d3" (UID: "3b56057d-1ca6-4fc5-a104-37634f5e19d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.015316 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-config" (OuterVolumeSpecName: "config") pod "3b56057d-1ca6-4fc5-a104-37634f5e19d3" (UID: "3b56057d-1ca6-4fc5-a104-37634f5e19d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.016022 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3b56057d-1ca6-4fc5-a104-37634f5e19d3" (UID: "3b56057d-1ca6-4fc5-a104-37634f5e19d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.016661 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b56057d-1ca6-4fc5-a104-37634f5e19d3-kube-api-access-tvt6t" (OuterVolumeSpecName: "kube-api-access-tvt6t") pod "3b56057d-1ca6-4fc5-a104-37634f5e19d3" (UID: "3b56057d-1ca6-4fc5-a104-37634f5e19d3"). InnerVolumeSpecName "kube-api-access-tvt6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.018735 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-combined-ca-bundle\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.023630 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-db-sync-config-data\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.032962 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdvk8\" (UniqueName: \"kubernetes.io/projected/bedfd09b-1fdb-43e9-88d7-47fd25103af6-kube-api-access-hdvk8\") pod \"barbican-db-sync-5765m\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.112645 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5765m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114009 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-config\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114085 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-certs\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114136 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzjgh\" (UniqueName: \"kubernetes.io/projected/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-kube-api-access-kzjgh\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114166 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114199 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-config-data\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnb2h\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-kube-api-access-lnb2h\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114374 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-scripts\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-combined-ca-bundle\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114443 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114479 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114557 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114577 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114591 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114605 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvt6t\" (UniqueName: \"kubernetes.io/projected/3b56057d-1ca6-4fc5-a104-37634f5e19d3-kube-api-access-tvt6t\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114621 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.114633 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b56057d-1ca6-4fc5-a104-37634f5e19d3-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.115541 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.117011 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.117819 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-config\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.118057 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.118508 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.151298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzjgh\" (UniqueName: \"kubernetes.io/projected/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-kube-api-access-kzjgh\") pod \"dnsmasq-dns-57c957c4ff-9tcn9\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.216193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-combined-ca-bundle\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.216304 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-certs\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.216352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-config-data\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.216400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnb2h\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-kube-api-access-lnb2h\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.216447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-scripts\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.219285 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8w9m9"] Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.227087 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-combined-ca-bundle\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.233068 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-certs\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.233489 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-scripts\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.235730 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-config-data\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.246127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnb2h\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-kube-api-access-lnb2h\") pod \"cloudkitty-db-sync-vprjr\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.251645 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.270789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.392887 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.402800 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.411912 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-r2c4m" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.412230 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.412394 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.474244 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.514089 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chfwn"] Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.552334 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.553820 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557277 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557360 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557401 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557420 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557461 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p98mt\" (UniqueName: \"kubernetes.io/projected/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-kube-api-access-p98mt\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557489 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.557515 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.558120 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.580077 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:24 crc kubenswrapper[4894]: W1210 19:16:24.598089 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod395f1d4d_ca65_4eb5_868a_e12ce3573151.slice/crio-95b785cef3fb3df8373745b3a0220514421ef92f0891b1d0d7f71504ebcfaf24 WatchSource:0}: Error finding container 95b785cef3fb3df8373745b3a0220514421ef92f0891b1d0d7f71504ebcfaf24: Status 404 returned error can't find the container with id 95b785cef3fb3df8373745b3a0220514421ef92f0891b1d0d7f71504ebcfaf24 Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.660652 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4b8x\" (UniqueName: \"kubernetes.io/projected/77f33b62-7220-492a-906b-cd619cb2f6dc-kube-api-access-r4b8x\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661062 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-logs\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661107 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661132 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661151 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661241 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661318 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p98mt\" (UniqueName: \"kubernetes.io/projected/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-kube-api-access-p98mt\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661336 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661405 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.661891 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.665298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.668147 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.673004 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.673049 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6b2877179eed34ff266ca183c106ee53ca8775eae08a086cf0505ade7a2914a4/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.687348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.690588 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.702060 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p98mt\" (UniqueName: \"kubernetes.io/projected/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-kube-api-access-p98mt\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.754607 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763179 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763235 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4b8x\" (UniqueName: \"kubernetes.io/projected/77f33b62-7220-492a-906b-cd619cb2f6dc-kube-api-access-r4b8x\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763264 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-logs\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763291 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763310 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.763341 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.764327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.764388 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-logs\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.765833 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.765888 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6b95216d4ff1b4536f5681f700c7b5d7fd226b07d8f87d718c34fcdf28bde918/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.767208 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.771864 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.772100 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.779999 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4b8x\" (UniqueName: \"kubernetes.io/projected/77f33b62-7220-492a-906b-cd619cb2f6dc-kube-api-access-r4b8x\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.806996 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.847256 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.869387 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" event={"ID":"395f1d4d-ca65-4eb5-868a-e12ce3573151","Type":"ContainerStarted","Data":"b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2"} Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.869430 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" event={"ID":"395f1d4d-ca65-4eb5-868a-e12ce3573151","Type":"ContainerStarted","Data":"95b785cef3fb3df8373745b3a0220514421ef92f0891b1d0d7f71504ebcfaf24"} Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.869547 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" podUID="395f1d4d-ca65-4eb5-868a-e12ce3573151" containerName="init" containerID="cri-o://b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2" gracePeriod=10 Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.891943 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerName="dnsmasq-dns" containerID="cri-o://6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32" gracePeriod=10 Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.892980 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8w9m9" event={"ID":"b1ad5751-b675-40b2-bae2-2e0de76257a3","Type":"ContainerStarted","Data":"3a3033834bef69f8782e1cff647ddcbc0d44dff33e2bb1c8d8f91898c281d0c9"} Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.893024 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8w9m9" event={"ID":"b1ad5751-b675-40b2-bae2-2e0de76257a3","Type":"ContainerStarted","Data":"b62dbab345d0aa63863f36dc302235ad1cc2b7467062ba8b69dc39d2b072b7a9"} Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.893058 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-hvdqp" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.899945 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.949120 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.953445 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8w9m9" podStartSLOduration=1.95342265 podStartE2EDuration="1.95342265s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:24.92721407 +0000 UTC m=+1267.722061862" watchObservedRunningTime="2025-12-10 19:16:24.95342265 +0000 UTC m=+1267.748270442" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.963156 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 10 19:16:24 crc kubenswrapper[4894]: I1210 19:16:24.973428 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.025656 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5765m"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.139895 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-hvdqp"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.150362 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-hvdqp"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.201529 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jk98n"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.240597 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5bq9r"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.268330 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4gmfn"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.329826 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.440628 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.498428 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-9tcn9"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.510334 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.516812 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.587133 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b56057d-1ca6-4fc5-a104-37634f5e19d3" path="/var/lib/kubelet/pods/3b56057d-1ca6-4fc5-a104-37634f5e19d3/volumes" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.599759 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-vprjr"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.616442 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrdxm\" (UniqueName: \"kubernetes.io/projected/395f1d4d-ca65-4eb5-868a-e12ce3573151-kube-api-access-lrdxm\") pod \"395f1d4d-ca65-4eb5-868a-e12ce3573151\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.619198 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-config\") pod \"395f1d4d-ca65-4eb5-868a-e12ce3573151\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.619306 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-swift-storage-0\") pod \"395f1d4d-ca65-4eb5-868a-e12ce3573151\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.619360 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-sb\") pod \"395f1d4d-ca65-4eb5-868a-e12ce3573151\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.619380 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-nb\") pod \"395f1d4d-ca65-4eb5-868a-e12ce3573151\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.619451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-svc\") pod \"395f1d4d-ca65-4eb5-868a-e12ce3573151\" (UID: \"395f1d4d-ca65-4eb5-868a-e12ce3573151\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.629737 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395f1d4d-ca65-4eb5-868a-e12ce3573151-kube-api-access-lrdxm" (OuterVolumeSpecName: "kube-api-access-lrdxm") pod "395f1d4d-ca65-4eb5-868a-e12ce3573151" (UID: "395f1d4d-ca65-4eb5-868a-e12ce3573151"). InnerVolumeSpecName "kube-api-access-lrdxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.659734 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "395f1d4d-ca65-4eb5-868a-e12ce3573151" (UID: "395f1d4d-ca65-4eb5-868a-e12ce3573151"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.659897 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "395f1d4d-ca65-4eb5-868a-e12ce3573151" (UID: "395f1d4d-ca65-4eb5-868a-e12ce3573151"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.662030 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-config" (OuterVolumeSpecName: "config") pod "395f1d4d-ca65-4eb5-868a-e12ce3573151" (UID: "395f1d4d-ca65-4eb5-868a-e12ce3573151"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.667310 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "395f1d4d-ca65-4eb5-868a-e12ce3573151" (UID: "395f1d4d-ca65-4eb5-868a-e12ce3573151"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.704050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "395f1d4d-ca65-4eb5-868a-e12ce3573151" (UID: "395f1d4d-ca65-4eb5-868a-e12ce3573151"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.722194 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.722230 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.722240 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.722248 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrdxm\" (UniqueName: \"kubernetes.io/projected/395f1d4d-ca65-4eb5-868a-e12ce3573151-kube-api-access-lrdxm\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.722258 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.722267 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/395f1d4d-ca65-4eb5-868a-e12ce3573151-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.744021 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.758594 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.827185 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-swift-storage-0\") pod \"b0348d25-e268-4a8d-a805-c3a33ce28b39\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.876114 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.929070 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-nb\") pod \"b0348d25-e268-4a8d-a805-c3a33ce28b39\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.929141 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpzxr\" (UniqueName: \"kubernetes.io/projected/b0348d25-e268-4a8d-a805-c3a33ce28b39-kube-api-access-wpzxr\") pod \"b0348d25-e268-4a8d-a805-c3a33ce28b39\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.929179 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-config\") pod \"b0348d25-e268-4a8d-a805-c3a33ce28b39\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.929199 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-svc\") pod \"b0348d25-e268-4a8d-a805-c3a33ce28b39\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.929252 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-sb\") pod \"b0348d25-e268-4a8d-a805-c3a33ce28b39\" (UID: \"b0348d25-e268-4a8d-a805-c3a33ce28b39\") " Dec 10 19:16:25 crc kubenswrapper[4894]: I1210 19:16:25.951125 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0348d25-e268-4a8d-a805-c3a33ce28b39-kube-api-access-wpzxr" (OuterVolumeSpecName: "kube-api-access-wpzxr") pod "b0348d25-e268-4a8d-a805-c3a33ce28b39" (UID: "b0348d25-e268-4a8d-a805-c3a33ce28b39"). InnerVolumeSpecName "kube-api-access-wpzxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.014871 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0348d25-e268-4a8d-a805-c3a33ce28b39" (UID: "b0348d25-e268-4a8d-a805-c3a33ce28b39"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.018931 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4gmfn" event={"ID":"7a781080-8482-4915-9d13-b88d1785f7f3","Type":"ContainerStarted","Data":"8382f482fbe41a5071517e5ecbb40b36f92fc2b058791eb194422a2fd5329539"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.033103 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerStarted","Data":"d1d39bc215c37245dedacce0112e5e2c43d0dbcf6f4b7ed36077a475467c8cb5"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.041562 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpzxr\" (UniqueName: \"kubernetes.io/projected/b0348d25-e268-4a8d-a805-c3a33ce28b39-kube-api-access-wpzxr\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.041591 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.047327 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b0348d25-e268-4a8d-a805-c3a33ce28b39" (UID: "b0348d25-e268-4a8d-a805-c3a33ce28b39"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.048673 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-config" (OuterVolumeSpecName: "config") pod "b0348d25-e268-4a8d-a805-c3a33ce28b39" (UID: "b0348d25-e268-4a8d-a805-c3a33ce28b39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.055100 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" event={"ID":"b6a2edcb-fcd4-41c4-8063-0afa95a99bda","Type":"ContainerStarted","Data":"0dfc3d2bd2c0ecd86660f6bad0e63e5c170d6ae70bcd1ccd5e63c61540542bab"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.057110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5765m" event={"ID":"bedfd09b-1fdb-43e9-88d7-47fd25103af6","Type":"ContainerStarted","Data":"1b1057939bc0c4515ff2c77aced64b39de3ccc0fbeddd3ff96adc6d6c7456a53"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.071027 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jk98n" event={"ID":"2a158cd9-dd00-4d21-bc0d-cf9f84aca910","Type":"ContainerStarted","Data":"dc891aa04e9236ae856c527ec99f570d500cf5af2ff64596115282fa028072e9"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.071062 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jk98n" event={"ID":"2a158cd9-dd00-4d21-bc0d-cf9f84aca910","Type":"ContainerStarted","Data":"d386d7a160a35ec146cf742fbc06decb43973cf156ffcec5c2af6aeb5f162fa6"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.087568 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0348d25-e268-4a8d-a805-c3a33ce28b39" (UID: "b0348d25-e268-4a8d-a805-c3a33ce28b39"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.121167 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vprjr" event={"ID":"bc3ed2c8-9e2c-443b-bc22-a39397986e13","Type":"ContainerStarted","Data":"a2cae11379f21211665785f92b7b4967e4857e5569d8c57526144881ac66c22d"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.123302 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5bq9r" event={"ID":"6524d3b3-56ee-4a03-84a5-dfa9c4c27926","Type":"ContainerStarted","Data":"7cae81252524e5eacb5c1bf72de9e6a5ba083cf357ce875e8503b651d7205bda"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.127913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8a604e1-9350-43e4-b5c0-a460ef9b8f99","Type":"ContainerStarted","Data":"e755b1f6e2b64e0ca2f050aeb5f854fc1e0afd659ec6addb5bfa6da0242421c0"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.129751 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jk98n" podStartSLOduration=3.129739738 podStartE2EDuration="3.129739738s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:26.123204924 +0000 UTC m=+1268.918052716" watchObservedRunningTime="2025-12-10 19:16:26.129739738 +0000 UTC m=+1268.924587550" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.136351 4894 generic.go:334] "Generic (PLEG): container finished" podID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerID="6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32" exitCode=0 Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.136415 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" event={"ID":"b0348d25-e268-4a8d-a805-c3a33ce28b39","Type":"ContainerDied","Data":"6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.136450 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" event={"ID":"b0348d25-e268-4a8d-a805-c3a33ce28b39","Type":"ContainerDied","Data":"2e9d965dc895a952119ac3bb39cbd4092726b1d652db0e27bb07f38cab82caaa"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.136470 4894 scope.go:117] "RemoveContainer" containerID="6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.136579 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-vz5cc" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.150086 4894 generic.go:334] "Generic (PLEG): container finished" podID="395f1d4d-ca65-4eb5-868a-e12ce3573151" containerID="b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2" exitCode=0 Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.150148 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" event={"ID":"395f1d4d-ca65-4eb5-868a-e12ce3573151","Type":"ContainerDied","Data":"b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.150383 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.151671 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.152898 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.153226 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.157611 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chfwn" event={"ID":"395f1d4d-ca65-4eb5-868a-e12ce3573151","Type":"ContainerDied","Data":"95b785cef3fb3df8373745b3a0220514421ef92f0891b1d0d7f71504ebcfaf24"} Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.164486 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.186674 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0348d25-e268-4a8d-a805-c3a33ce28b39" (UID: "b0348d25-e268-4a8d-a805-c3a33ce28b39"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.257635 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0348d25-e268-4a8d-a805-c3a33ce28b39-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.335766 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chfwn"] Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.345212 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chfwn"] Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.369026 4894 scope.go:117] "RemoveContainer" containerID="efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.470726 4894 scope.go:117] "RemoveContainer" containerID="6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32" Dec 10 19:16:26 crc kubenswrapper[4894]: E1210 19:16:26.476076 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32\": container with ID starting with 6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32 not found: ID does not exist" containerID="6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.476125 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32"} err="failed to get container status \"6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32\": rpc error: code = NotFound desc = could not find container \"6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32\": container with ID starting with 6822d815b7e926a23d3a15d335a0621f1099f2115fa24fdae78e1e620bbf4a32 not found: ID does not exist" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.476153 4894 scope.go:117] "RemoveContainer" containerID="efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0" Dec 10 19:16:26 crc kubenswrapper[4894]: E1210 19:16:26.497062 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0\": container with ID starting with efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0 not found: ID does not exist" containerID="efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.497120 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0"} err="failed to get container status \"efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0\": rpc error: code = NotFound desc = could not find container \"efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0\": container with ID starting with efcad097c737c6fc5de27c327817cc1cc97e38a3c55f52f596c315b397414dc0 not found: ID does not exist" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.497154 4894 scope.go:117] "RemoveContainer" containerID="b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.576561 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-vz5cc"] Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.582495 4894 scope.go:117] "RemoveContainer" containerID="b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.582574 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-vz5cc"] Dec 10 19:16:26 crc kubenswrapper[4894]: E1210 19:16:26.585519 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2\": container with ID starting with b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2 not found: ID does not exist" containerID="b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2" Dec 10 19:16:26 crc kubenswrapper[4894]: I1210 19:16:26.585563 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2"} err="failed to get container status \"b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2\": rpc error: code = NotFound desc = could not find container \"b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2\": container with ID starting with b72a0f38d1d1bc3d10a20431ed9c6684dbc53b180adc1e27de54f9010a0875c2 not found: ID does not exist" Dec 10 19:16:26 crc kubenswrapper[4894]: E1210 19:16:26.775293 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0348d25_e268_4a8d_a805_c3a33ce28b39.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.166919 4894 generic.go:334] "Generic (PLEG): container finished" podID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerID="0de6da4072a02827caeeb10d6384203d342b373f0a395dd2915d18044ca9806b" exitCode=0 Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.167002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" event={"ID":"b6a2edcb-fcd4-41c4-8063-0afa95a99bda","Type":"ContainerDied","Data":"0de6da4072a02827caeeb10d6384203d342b373f0a395dd2915d18044ca9806b"} Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.167414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" event={"ID":"b6a2edcb-fcd4-41c4-8063-0afa95a99bda","Type":"ContainerStarted","Data":"704c9a990ba89ff9bc86b6cec93f3ac4880b14bdb45ea31107ce5d32d2eaf7e7"} Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.174149 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.174810 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77f33b62-7220-492a-906b-cd619cb2f6dc","Type":"ContainerStarted","Data":"8b013801a3664d09e25c4fbf952542b0a1c3a55449fd4d6c09ab36eb6982a920"} Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.178964 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8a604e1-9350-43e4-b5c0-a460ef9b8f99","Type":"ContainerStarted","Data":"720e74e5c3037706d3a033689ded26b6a8821352254887b062eb709b54ef1134"} Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.218999 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" podStartSLOduration=4.218979603 podStartE2EDuration="4.218979603s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:27.217580286 +0000 UTC m=+1270.012428088" watchObservedRunningTime="2025-12-10 19:16:27.218979603 +0000 UTC m=+1270.013827395" Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.566947 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395f1d4d-ca65-4eb5-868a-e12ce3573151" path="/var/lib/kubelet/pods/395f1d4d-ca65-4eb5-868a-e12ce3573151/volumes" Dec 10 19:16:27 crc kubenswrapper[4894]: I1210 19:16:27.568079 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" path="/var/lib/kubelet/pods/b0348d25-e268-4a8d-a805-c3a33ce28b39/volumes" Dec 10 19:16:28 crc kubenswrapper[4894]: I1210 19:16:28.242473 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77f33b62-7220-492a-906b-cd619cb2f6dc","Type":"ContainerStarted","Data":"db3858f70c2373f7b9d167314d7bb6a4dc8f7183fe008e529bb1cd29c6d2dc8c"} Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.256124 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77f33b62-7220-492a-906b-cd619cb2f6dc","Type":"ContainerStarted","Data":"5fbf9c6f6e2cf908f1d28cc1dcfc09d639506bc23c8e3a60928ef81aa7085e4b"} Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.256192 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-log" containerID="cri-o://db3858f70c2373f7b9d167314d7bb6a4dc8f7183fe008e529bb1cd29c6d2dc8c" gracePeriod=30 Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.256420 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-httpd" containerID="cri-o://5fbf9c6f6e2cf908f1d28cc1dcfc09d639506bc23c8e3a60928ef81aa7085e4b" gracePeriod=30 Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.260792 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8a604e1-9350-43e4-b5c0-a460ef9b8f99","Type":"ContainerStarted","Data":"160692ec032da12f7ebd10bd415c9a50c5facaff83f67ec6b295ae2bcd32ca90"} Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.260982 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-log" containerID="cri-o://720e74e5c3037706d3a033689ded26b6a8821352254887b062eb709b54ef1134" gracePeriod=30 Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.261087 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-httpd" containerID="cri-o://160692ec032da12f7ebd10bd415c9a50c5facaff83f67ec6b295ae2bcd32ca90" gracePeriod=30 Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.321616 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.321598021 podStartE2EDuration="6.321598021s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:29.287187283 +0000 UTC m=+1272.082035135" watchObservedRunningTime="2025-12-10 19:16:29.321598021 +0000 UTC m=+1272.116445813" Dec 10 19:16:29 crc kubenswrapper[4894]: I1210 19:16:29.323216 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.323206194 podStartE2EDuration="6.323206194s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:16:29.315433667 +0000 UTC m=+1272.110281489" watchObservedRunningTime="2025-12-10 19:16:29.323206194 +0000 UTC m=+1272.118053986" Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.278038 4894 generic.go:334] "Generic (PLEG): container finished" podID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerID="5fbf9c6f6e2cf908f1d28cc1dcfc09d639506bc23c8e3a60928ef81aa7085e4b" exitCode=0 Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.278074 4894 generic.go:334] "Generic (PLEG): container finished" podID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerID="db3858f70c2373f7b9d167314d7bb6a4dc8f7183fe008e529bb1cd29c6d2dc8c" exitCode=143 Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.278077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77f33b62-7220-492a-906b-cd619cb2f6dc","Type":"ContainerDied","Data":"5fbf9c6f6e2cf908f1d28cc1dcfc09d639506bc23c8e3a60928ef81aa7085e4b"} Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.278129 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77f33b62-7220-492a-906b-cd619cb2f6dc","Type":"ContainerDied","Data":"db3858f70c2373f7b9d167314d7bb6a4dc8f7183fe008e529bb1cd29c6d2dc8c"} Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.282765 4894 generic.go:334] "Generic (PLEG): container finished" podID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerID="160692ec032da12f7ebd10bd415c9a50c5facaff83f67ec6b295ae2bcd32ca90" exitCode=0 Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.282790 4894 generic.go:334] "Generic (PLEG): container finished" podID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerID="720e74e5c3037706d3a033689ded26b6a8821352254887b062eb709b54ef1134" exitCode=143 Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.282812 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8a604e1-9350-43e4-b5c0-a460ef9b8f99","Type":"ContainerDied","Data":"160692ec032da12f7ebd10bd415c9a50c5facaff83f67ec6b295ae2bcd32ca90"} Dec 10 19:16:30 crc kubenswrapper[4894]: I1210 19:16:30.282838 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8a604e1-9350-43e4-b5c0-a460ef9b8f99","Type":"ContainerDied","Data":"720e74e5c3037706d3a033689ded26b6a8821352254887b062eb709b54ef1134"} Dec 10 19:16:31 crc kubenswrapper[4894]: I1210 19:16:31.304775 4894 generic.go:334] "Generic (PLEG): container finished" podID="b1ad5751-b675-40b2-bae2-2e0de76257a3" containerID="3a3033834bef69f8782e1cff647ddcbc0d44dff33e2bb1c8d8f91898c281d0c9" exitCode=0 Dec 10 19:16:31 crc kubenswrapper[4894]: I1210 19:16:31.304884 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8w9m9" event={"ID":"b1ad5751-b675-40b2-bae2-2e0de76257a3","Type":"ContainerDied","Data":"3a3033834bef69f8782e1cff647ddcbc0d44dff33e2bb1c8d8f91898c281d0c9"} Dec 10 19:16:34 crc kubenswrapper[4894]: I1210 19:16:34.254001 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:16:34 crc kubenswrapper[4894]: I1210 19:16:34.368282 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5dlsv"] Dec 10 19:16:34 crc kubenswrapper[4894]: I1210 19:16:34.368504 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" containerID="cri-o://3eac62ee0066a6901bf1400403d95b9e486a4c214e4f28d0b09fe699eacb2646" gracePeriod=10 Dec 10 19:16:35 crc kubenswrapper[4894]: I1210 19:16:35.405993 4894 generic.go:334] "Generic (PLEG): container finished" podID="115d719b-0a27-44a4-b117-a20211e65090" containerID="3eac62ee0066a6901bf1400403d95b9e486a4c214e4f28d0b09fe699eacb2646" exitCode=0 Dec 10 19:16:35 crc kubenswrapper[4894]: I1210 19:16:35.406030 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" event={"ID":"115d719b-0a27-44a4-b117-a20211e65090","Type":"ContainerDied","Data":"3eac62ee0066a6901bf1400403d95b9e486a4c214e4f28d0b09fe699eacb2646"} Dec 10 19:16:36 crc kubenswrapper[4894]: I1210 19:16:36.480588 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.423462 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.433406 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.444755 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8w9m9" event={"ID":"b1ad5751-b675-40b2-bae2-2e0de76257a3","Type":"ContainerDied","Data":"b62dbab345d0aa63863f36dc302235ad1cc2b7467062ba8b69dc39d2b072b7a9"} Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.444794 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b62dbab345d0aa63863f36dc302235ad1cc2b7467062ba8b69dc39d2b072b7a9" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.444873 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8w9m9" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.449256 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77f33b62-7220-492a-906b-cd619cb2f6dc","Type":"ContainerDied","Data":"8b013801a3664d09e25c4fbf952542b0a1c3a55449fd4d6c09ab36eb6982a920"} Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.449310 4894 scope.go:117] "RemoveContainer" containerID="5fbf9c6f6e2cf908f1d28cc1dcfc09d639506bc23c8e3a60928ef81aa7085e4b" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.449440 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499735 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-config-data\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499781 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-config-data\") pod \"b1ad5751-b675-40b2-bae2-2e0de76257a3\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499809 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-logs\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499887 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-httpd-run\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499917 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-scripts\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499931 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-scripts\") pod \"b1ad5751-b675-40b2-bae2-2e0de76257a3\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.499952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4b8x\" (UniqueName: \"kubernetes.io/projected/77f33b62-7220-492a-906b-cd619cb2f6dc-kube-api-access-r4b8x\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.500066 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.500111 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4vtm\" (UniqueName: \"kubernetes.io/projected/b1ad5751-b675-40b2-bae2-2e0de76257a3-kube-api-access-r4vtm\") pod \"b1ad5751-b675-40b2-bae2-2e0de76257a3\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.500190 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-combined-ca-bundle\") pod \"77f33b62-7220-492a-906b-cd619cb2f6dc\" (UID: \"77f33b62-7220-492a-906b-cd619cb2f6dc\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.500212 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-combined-ca-bundle\") pod \"b1ad5751-b675-40b2-bae2-2e0de76257a3\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.500248 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-fernet-keys\") pod \"b1ad5751-b675-40b2-bae2-2e0de76257a3\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.500286 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-credential-keys\") pod \"b1ad5751-b675-40b2-bae2-2e0de76257a3\" (UID: \"b1ad5751-b675-40b2-bae2-2e0de76257a3\") " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.506633 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-scripts" (OuterVolumeSpecName: "scripts") pod "b1ad5751-b675-40b2-bae2-2e0de76257a3" (UID: "b1ad5751-b675-40b2-bae2-2e0de76257a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.507027 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b1ad5751-b675-40b2-bae2-2e0de76257a3" (UID: "b1ad5751-b675-40b2-bae2-2e0de76257a3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.510086 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-logs" (OuterVolumeSpecName: "logs") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.510333 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f33b62-7220-492a-906b-cd619cb2f6dc-kube-api-access-r4b8x" (OuterVolumeSpecName: "kube-api-access-r4b8x") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "kube-api-access-r4b8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.511216 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.516084 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b1ad5751-b675-40b2-bae2-2e0de76257a3" (UID: "b1ad5751-b675-40b2-bae2-2e0de76257a3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.520291 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ad5751-b675-40b2-bae2-2e0de76257a3-kube-api-access-r4vtm" (OuterVolumeSpecName: "kube-api-access-r4vtm") pod "b1ad5751-b675-40b2-bae2-2e0de76257a3" (UID: "b1ad5751-b675-40b2-bae2-2e0de76257a3"). InnerVolumeSpecName "kube-api-access-r4vtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.536505 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5" (OuterVolumeSpecName: "glance") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.537800 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-scripts" (OuterVolumeSpecName: "scripts") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.559726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-config-data" (OuterVolumeSpecName: "config-data") pod "b1ad5751-b675-40b2-bae2-2e0de76257a3" (UID: "b1ad5751-b675-40b2-bae2-2e0de76257a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.574004 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1ad5751-b675-40b2-bae2-2e0de76257a3" (UID: "b1ad5751-b675-40b2-bae2-2e0de76257a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.593018 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.595940 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-config-data" (OuterVolumeSpecName: "config-data") pod "77f33b62-7220-492a-906b-cd619cb2f6dc" (UID: "77f33b62-7220-492a-906b-cd619cb2f6dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602820 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602857 4894 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602868 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602877 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602885 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602895 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77f33b62-7220-492a-906b-cd619cb2f6dc-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602904 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602911 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602919 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4b8x\" (UniqueName: \"kubernetes.io/projected/77f33b62-7220-492a-906b-cd619cb2f6dc-kube-api-access-r4b8x\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602939 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") on node \"crc\" " Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602950 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4vtm\" (UniqueName: \"kubernetes.io/projected/b1ad5751-b675-40b2-bae2-2e0de76257a3-kube-api-access-r4vtm\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602959 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f33b62-7220-492a-906b-cd619cb2f6dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.602967 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ad5751-b675-40b2-bae2-2e0de76257a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.649918 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.650058 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5") on node "crc" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.704219 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.789253 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.806026 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.816904 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:38 crc kubenswrapper[4894]: E1210 19:16:38.817351 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ad5751-b675-40b2-bae2-2e0de76257a3" containerName="keystone-bootstrap" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817370 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ad5751-b675-40b2-bae2-2e0de76257a3" containerName="keystone-bootstrap" Dec 10 19:16:38 crc kubenswrapper[4894]: E1210 19:16:38.817389 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerName="dnsmasq-dns" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817396 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerName="dnsmasq-dns" Dec 10 19:16:38 crc kubenswrapper[4894]: E1210 19:16:38.817414 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395f1d4d-ca65-4eb5-868a-e12ce3573151" containerName="init" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817422 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="395f1d4d-ca65-4eb5-868a-e12ce3573151" containerName="init" Dec 10 19:16:38 crc kubenswrapper[4894]: E1210 19:16:38.817442 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-log" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817449 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-log" Dec 10 19:16:38 crc kubenswrapper[4894]: E1210 19:16:38.817459 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerName="init" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817465 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerName="init" Dec 10 19:16:38 crc kubenswrapper[4894]: E1210 19:16:38.817477 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-httpd" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817482 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-httpd" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817652 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ad5751-b675-40b2-bae2-2e0de76257a3" containerName="keystone-bootstrap" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817670 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0348d25-e268-4a8d-a805-c3a33ce28b39" containerName="dnsmasq-dns" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817681 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="395f1d4d-ca65-4eb5-868a-e12ce3573151" containerName="init" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817696 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-httpd" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.817710 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" containerName="glance-log" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.818731 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.825050 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.854521 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.854741 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908296 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-config-data\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908362 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908404 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908448 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-logs\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908496 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z28xd\" (UniqueName: \"kubernetes.io/projected/9874ce74-c8e9-4719-802b-81a69271131a-kube-api-access-z28xd\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908542 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908656 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-scripts\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:38 crc kubenswrapper[4894]: I1210 19:16:38.908694 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.010991 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-config-data\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.011417 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.011626 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.012074 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-logs\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.012233 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z28xd\" (UniqueName: \"kubernetes.io/projected/9874ce74-c8e9-4719-802b-81a69271131a-kube-api-access-z28xd\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.012448 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.012613 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-logs\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.013242 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-scripts\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.013730 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.013535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.016780 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.016939 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-scripts\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.017028 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.017286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-config-data\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.020559 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.020585 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6b95216d4ff1b4536f5681f700c7b5d7fd226b07d8f87d718c34fcdf28bde918/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.030611 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z28xd\" (UniqueName: \"kubernetes.io/projected/9874ce74-c8e9-4719-802b-81a69271131a-kube-api-access-z28xd\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.060873 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.169471 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.556143 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f33b62-7220-492a-906b-cd619cb2f6dc" path="/var/lib/kubelet/pods/77f33b62-7220-492a-906b-cd619cb2f6dc/volumes" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.583910 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8w9m9"] Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.592348 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8w9m9"] Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.689211 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b2w4f"] Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.691001 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.692962 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ph6mx" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.693214 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.693267 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.693723 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.703611 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b2w4f"] Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.733709 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tx5t\" (UniqueName: \"kubernetes.io/projected/8aac2573-2b49-4dcf-b808-d1adc6435bc9-kube-api-access-4tx5t\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.733799 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-config-data\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.733906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-combined-ca-bundle\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.733931 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-fernet-keys\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.733994 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-credential-keys\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.734041 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-scripts\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.835233 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tx5t\" (UniqueName: \"kubernetes.io/projected/8aac2573-2b49-4dcf-b808-d1adc6435bc9-kube-api-access-4tx5t\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.835309 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-config-data\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.835393 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-combined-ca-bundle\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.835411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-fernet-keys\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.835456 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-credential-keys\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.835492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-scripts\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.840302 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-config-data\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.840557 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-scripts\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.841007 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-combined-ca-bundle\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.849379 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-credential-keys\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.860575 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tx5t\" (UniqueName: \"kubernetes.io/projected/8aac2573-2b49-4dcf-b808-d1adc6435bc9-kube-api-access-4tx5t\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:39 crc kubenswrapper[4894]: I1210 19:16:39.860577 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-fernet-keys\") pod \"keystone-bootstrap-b2w4f\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:40 crc kubenswrapper[4894]: I1210 19:16:40.023317 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:16:41 crc kubenswrapper[4894]: I1210 19:16:41.553648 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1ad5751-b675-40b2-bae2-2e0de76257a3" path="/var/lib/kubelet/pods/b1ad5751-b675-40b2-bae2-2e0de76257a3/volumes" Dec 10 19:16:46 crc kubenswrapper[4894]: I1210 19:16:46.479945 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 10 19:16:49 crc kubenswrapper[4894]: I1210 19:16:49.554477 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a158cd9-dd00-4d21-bc0d-cf9f84aca910" containerID="dc891aa04e9236ae856c527ec99f570d500cf5af2ff64596115282fa028072e9" exitCode=0 Dec 10 19:16:49 crc kubenswrapper[4894]: I1210 19:16:49.554569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jk98n" event={"ID":"2a158cd9-dd00-4d21-bc0d-cf9f84aca910","Type":"ContainerDied","Data":"dc891aa04e9236ae856c527ec99f570d500cf5af2ff64596115282fa028072e9"} Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.399397 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.490898 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.490990 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.599010 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-sb\") pod \"115d719b-0a27-44a4-b117-a20211e65090\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.599079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn664\" (UniqueName: \"kubernetes.io/projected/115d719b-0a27-44a4-b117-a20211e65090-kube-api-access-sn664\") pod \"115d719b-0a27-44a4-b117-a20211e65090\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.599124 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-nb\") pod \"115d719b-0a27-44a4-b117-a20211e65090\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.599169 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-dns-svc\") pod \"115d719b-0a27-44a4-b117-a20211e65090\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.599201 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-config\") pod \"115d719b-0a27-44a4-b117-a20211e65090\" (UID: \"115d719b-0a27-44a4-b117-a20211e65090\") " Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.604860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/115d719b-0a27-44a4-b117-a20211e65090-kube-api-access-sn664" (OuterVolumeSpecName: "kube-api-access-sn664") pod "115d719b-0a27-44a4-b117-a20211e65090" (UID: "115d719b-0a27-44a4-b117-a20211e65090"). InnerVolumeSpecName "kube-api-access-sn664". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.674507 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" event={"ID":"115d719b-0a27-44a4-b117-a20211e65090","Type":"ContainerDied","Data":"2fc020f0584646bb06af9f8ba60a0641f7c1377bce06d1fddb18e71cde3feacd"} Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.674613 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5dlsv" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.693247 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "115d719b-0a27-44a4-b117-a20211e65090" (UID: "115d719b-0a27-44a4-b117-a20211e65090"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.704178 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.704211 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn664\" (UniqueName: \"kubernetes.io/projected/115d719b-0a27-44a4-b117-a20211e65090-kube-api-access-sn664\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.725964 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "115d719b-0a27-44a4-b117-a20211e65090" (UID: "115d719b-0a27-44a4-b117-a20211e65090"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.763373 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-config" (OuterVolumeSpecName: "config") pod "115d719b-0a27-44a4-b117-a20211e65090" (UID: "115d719b-0a27-44a4-b117-a20211e65090"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.767419 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "115d719b-0a27-44a4-b117-a20211e65090" (UID: "115d719b-0a27-44a4-b117-a20211e65090"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.812792 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.812833 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:51 crc kubenswrapper[4894]: I1210 19:16:51.812846 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/115d719b-0a27-44a4-b117-a20211e65090-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.016946 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5dlsv"] Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.025801 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5dlsv"] Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.055165 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.055349 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hdvk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-5765m_openstack(bedfd09b-1fdb-43e9-88d7-47fd25103af6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.056849 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-5765m" podUID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.064567 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.067693 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219372 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-scripts\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p98mt\" (UniqueName: \"kubernetes.io/projected/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-kube-api-access-p98mt\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219628 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npns8\" (UniqueName: \"kubernetes.io/projected/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-kube-api-access-npns8\") pod \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219803 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-combined-ca-bundle\") pod \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219882 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-combined-ca-bundle\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219943 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-httpd-run\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.219993 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-config-data\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.220023 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-config\") pod \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\" (UID: \"2a158cd9-dd00-4d21-bc0d-cf9f84aca910\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.220138 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-logs\") pod \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\" (UID: \"c8a604e1-9350-43e4-b5c0-a460ef9b8f99\") " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.221114 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-logs" (OuterVolumeSpecName: "logs") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.221394 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.223639 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-kube-api-access-npns8" (OuterVolumeSpecName: "kube-api-access-npns8") pod "2a158cd9-dd00-4d21-bc0d-cf9f84aca910" (UID: "2a158cd9-dd00-4d21-bc0d-cf9f84aca910"). InnerVolumeSpecName "kube-api-access-npns8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.224507 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-scripts" (OuterVolumeSpecName: "scripts") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.230621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-kube-api-access-p98mt" (OuterVolumeSpecName: "kube-api-access-p98mt") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "kube-api-access-p98mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.238511 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4" (OuterVolumeSpecName: "glance") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "pvc-102f2a85-63ab-4ac5-99ed-6047367290b4". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.245054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.245542 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a158cd9-dd00-4d21-bc0d-cf9f84aca910" (UID: "2a158cd9-dd00-4d21-bc0d-cf9f84aca910"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.252563 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-config" (OuterVolumeSpecName: "config") pod "2a158cd9-dd00-4d21-bc0d-cf9f84aca910" (UID: "2a158cd9-dd00-4d21-bc0d-cf9f84aca910"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.270533 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-config-data" (OuterVolumeSpecName: "config-data") pod "c8a604e1-9350-43e4-b5c0-a460ef9b8f99" (UID: "c8a604e1-9350-43e4-b5c0-a460ef9b8f99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326774 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p98mt\" (UniqueName: \"kubernetes.io/projected/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-kube-api-access-p98mt\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326868 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") on node \"crc\" " Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326887 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npns8\" (UniqueName: \"kubernetes.io/projected/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-kube-api-access-npns8\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326902 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326914 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326926 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326938 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326949 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a158cd9-dd00-4d21-bc0d-cf9f84aca910-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326960 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.326973 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8a604e1-9350-43e4-b5c0-a460ef9b8f99-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.353803 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.354029 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-102f2a85-63ab-4ac5-99ed-6047367290b4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4") on node "crc" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.428387 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") on node \"crc\" DevicePath \"\"" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.690131 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.690166 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c8a604e1-9350-43e4-b5c0-a460ef9b8f99","Type":"ContainerDied","Data":"e755b1f6e2b64e0ca2f050aeb5f854fc1e0afd659ec6addb5bfa6da0242421c0"} Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.710774 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jk98n" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.713097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jk98n" event={"ID":"2a158cd9-dd00-4d21-bc0d-cf9f84aca910","Type":"ContainerDied","Data":"d386d7a160a35ec146cf742fbc06decb43973cf156ffcec5c2af6aeb5f162fa6"} Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.713195 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d386d7a160a35ec146cf742fbc06decb43973cf156ffcec5c2af6aeb5f162fa6" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.713972 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-5765m" podUID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.742295 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.756585 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767299 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.767637 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-log" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767648 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-log" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.767660 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a158cd9-dd00-4d21-bc0d-cf9f84aca910" containerName="neutron-db-sync" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767665 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a158cd9-dd00-4d21-bc0d-cf9f84aca910" containerName="neutron-db-sync" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.767672 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-httpd" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767678 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-httpd" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.767688 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767694 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" Dec 10 19:16:52 crc kubenswrapper[4894]: E1210 19:16:52.767716 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="init" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767721 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="init" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767894 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a158cd9-dd00-4d21-bc0d-cf9f84aca910" containerName="neutron-db-sync" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767909 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-httpd" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767920 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="115d719b-0a27-44a4-b117-a20211e65090" containerName="dnsmasq-dns" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.767931 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" containerName="glance-log" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.768883 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.770920 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.771910 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.785036 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.937909 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.937991 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.938053 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.938191 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-logs\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.938246 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.938284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsdv5\" (UniqueName: \"kubernetes.io/projected/ca7b0207-4a62-4299-aa2a-14244c283866-kube-api-access-qsdv5\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.938322 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:52 crc kubenswrapper[4894]: I1210 19:16:52.938379 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.039677 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-logs\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040105 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsdv5\" (UniqueName: \"kubernetes.io/projected/ca7b0207-4a62-4299-aa2a-14244c283866-kube-api-access-qsdv5\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040207 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040523 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040766 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040915 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.040380 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-logs\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.042118 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.047029 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.047338 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.047439 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6b2877179eed34ff266ca183c106ee53ca8775eae08a086cf0505ade7a2914a4/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.055081 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.056376 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.057148 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.057288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsdv5\" (UniqueName: \"kubernetes.io/projected/ca7b0207-4a62-4299-aa2a-14244c283866-kube-api-access-qsdv5\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.100397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.371761 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-nct88"] Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.377065 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.399655 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.402208 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-nct88"] Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.504477 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8659fd4574-zvfv8"] Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.506947 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.511682 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.512032 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.512157 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.512287 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-72ww2" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.517230 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8659fd4574-zvfv8"] Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.553073 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.553187 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fdzw\" (UniqueName: \"kubernetes.io/projected/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-kube-api-access-8fdzw\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.553310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.553342 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.553375 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.553524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-config\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.567389 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="115d719b-0a27-44a4-b117-a20211e65090" path="/var/lib/kubelet/pods/115d719b-0a27-44a4-b117-a20211e65090/volumes" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.570256 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8a604e1-9350-43e4-b5c0-a460ef9b8f99" path="/var/lib/kubelet/pods/c8a604e1-9350-43e4-b5c0-a460ef9b8f99/volumes" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655058 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fdzw\" (UniqueName: \"kubernetes.io/projected/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-kube-api-access-8fdzw\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655317 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-combined-ca-bundle\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655411 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-config\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655601 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655837 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.655991 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-config\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.656139 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7lb7\" (UniqueName: \"kubernetes.io/projected/aadf98aa-0c6c-4d24-a847-e016114e615c-kube-api-access-q7lb7\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.656267 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.656559 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-httpd-config\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.656916 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-ovndb-tls-certs\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.657543 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.657953 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.658140 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-config\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.658437 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.658560 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.675835 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fdzw\" (UniqueName: \"kubernetes.io/projected/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-kube-api-access-8fdzw\") pod \"dnsmasq-dns-5ccc5c4795-nct88\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.722576 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.759053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-combined-ca-bundle\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.759093 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-config\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.759190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7lb7\" (UniqueName: \"kubernetes.io/projected/aadf98aa-0c6c-4d24-a847-e016114e615c-kube-api-access-q7lb7\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.759232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-httpd-config\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.759257 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-ovndb-tls-certs\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.763166 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-ovndb-tls-certs\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.763879 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-combined-ca-bundle\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.764447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-httpd-config\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.766143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-config\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.777131 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7lb7\" (UniqueName: \"kubernetes.io/projected/aadf98aa-0c6c-4d24-a847-e016114e615c-kube-api-access-q7lb7\") pod \"neutron-8659fd4574-zvfv8\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:53 crc kubenswrapper[4894]: I1210 19:16:53.829573 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:16:54 crc kubenswrapper[4894]: E1210 19:16:54.179000 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 10 19:16:54 crc kubenswrapper[4894]: E1210 19:16:54.179463 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gcdch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-5bq9r_openstack(6524d3b3-56ee-4a03-84a5-dfa9c4c27926): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4894]: E1210 19:16:54.180710 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-5bq9r" podUID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" Dec 10 19:16:54 crc kubenswrapper[4894]: I1210 19:16:54.185108 4894 scope.go:117] "RemoveContainer" containerID="db3858f70c2373f7b9d167314d7bb6a4dc8f7183fe008e529bb1cd29c6d2dc8c" Dec 10 19:16:54 crc kubenswrapper[4894]: E1210 19:16:54.747410 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-5bq9r" podUID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.432894 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54d58d7797-fnl7d"] Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.437458 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.439534 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.440181 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.466012 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54d58d7797-fnl7d"] Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.599672 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-ovndb-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.599720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-config\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.599936 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-public-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.600060 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-internal-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.600089 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmqgf\" (UniqueName: \"kubernetes.io/projected/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-kube-api-access-cmqgf\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.600386 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-httpd-config\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.600461 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-combined-ca-bundle\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702184 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-public-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702245 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-internal-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmqgf\" (UniqueName: \"kubernetes.io/projected/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-kube-api-access-cmqgf\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702302 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-httpd-config\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-combined-ca-bundle\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-ovndb-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.702425 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-config\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.708550 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-ovndb-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.709164 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-httpd-config\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.710739 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-internal-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.712965 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-config\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.715516 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-public-tls-certs\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.720379 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmqgf\" (UniqueName: \"kubernetes.io/projected/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-kube-api-access-cmqgf\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.720671 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8-combined-ca-bundle\") pod \"neutron-54d58d7797-fnl7d\" (UID: \"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8\") " pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:16:55 crc kubenswrapper[4894]: I1210 19:16:55.800581 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:17:00 crc kubenswrapper[4894]: I1210 19:17:00.560959 4894 scope.go:117] "RemoveContainer" containerID="3eac62ee0066a6901bf1400403d95b9e486a4c214e4f28d0b09fe699eacb2646" Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.047036 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b2w4f"] Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.686975 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.724241 4894 scope.go:117] "RemoveContainer" containerID="e680f26706f00ac782aefd533381d1fbdff67881cf2e4d3812e5e07eacaf1615" Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.872812 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b2w4f" event={"ID":"8aac2573-2b49-4dcf-b808-d1adc6435bc9","Type":"ContainerStarted","Data":"168dd2a6938a264a69ae1d2ffdf9c1546b577b14f80d5e331c0260aa9b905d09"} Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.907896 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9874ce74-c8e9-4719-802b-81a69271131a","Type":"ContainerStarted","Data":"ad8514fd3146c036a58e953aaf9e513597474c429a6b20ca325e87aff4f4bca0"} Dec 10 19:17:01 crc kubenswrapper[4894]: E1210 19:17:01.945820 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Dec 10 19:17:01 crc kubenswrapper[4894]: E1210 19:17:01.945902 4894 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Dec 10 19:17:01 crc kubenswrapper[4894]: E1210 19:17:01.946052 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lnb2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-vprjr_openstack(bc3ed2c8-9e2c-443b-bc22-a39397986e13): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:17:01 crc kubenswrapper[4894]: E1210 19:17:01.947244 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-vprjr" podUID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.950285 4894 scope.go:117] "RemoveContainer" containerID="160692ec032da12f7ebd10bd415c9a50c5facaff83f67ec6b295ae2bcd32ca90" Dec 10 19:17:01 crc kubenswrapper[4894]: I1210 19:17:01.995597 4894 scope.go:117] "RemoveContainer" containerID="720e74e5c3037706d3a033689ded26b6a8821352254887b062eb709b54ef1134" Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.216472 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-nct88"] Dec 10 19:17:02 crc kubenswrapper[4894]: W1210 19:17:02.230795 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7de86a55_00ea_44a6_aa0e_a9687e4fd56e.slice/crio-9a34d78b6f4660824d19f02e3ae0ca712d353b9658d28cab6615690b6d24eede WatchSource:0}: Error finding container 9a34d78b6f4660824d19f02e3ae0ca712d353b9658d28cab6615690b6d24eede: Status 404 returned error can't find the container with id 9a34d78b6f4660824d19f02e3ae0ca712d353b9658d28cab6615690b6d24eede Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.474710 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:17:02 crc kubenswrapper[4894]: W1210 19:17:02.493897 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca7b0207_4a62_4299_aa2a_14244c283866.slice/crio-d68d176036272d46ccf0d6f0e4b06085fdc4b61e9b215dbf65fee0e64032398e WatchSource:0}: Error finding container d68d176036272d46ccf0d6f0e4b06085fdc4b61e9b215dbf65fee0e64032398e: Status 404 returned error can't find the container with id d68d176036272d46ccf0d6f0e4b06085fdc4b61e9b215dbf65fee0e64032398e Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.619058 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54d58d7797-fnl7d"] Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.918408 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4gmfn" event={"ID":"7a781080-8482-4915-9d13-b88d1785f7f3","Type":"ContainerStarted","Data":"6eeddc13ccc3804f8d502f50d581d43f84041df7a038c0e510a577216b5c622b"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.920065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b2w4f" event={"ID":"8aac2573-2b49-4dcf-b808-d1adc6435bc9","Type":"ContainerStarted","Data":"199379249cc0934c4aaa80bdb58285e0403aecb4e7f339830f8d02331bf82029"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.923041 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9874ce74-c8e9-4719-802b-81a69271131a","Type":"ContainerStarted","Data":"592ae354c14bfc45e128bbf3ae8dcbbc9005608ae52f5d15b85262957b401560"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.926898 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ca7b0207-4a62-4299-aa2a-14244c283866","Type":"ContainerStarted","Data":"d68d176036272d46ccf0d6f0e4b06085fdc4b61e9b215dbf65fee0e64032398e"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.928441 4894 generic.go:334] "Generic (PLEG): container finished" podID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerID="67185d51fb7171ecf915a0e5cc1444cb49c59f5f6bf106aa30de576a4f77fd7f" exitCode=0 Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.928933 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" event={"ID":"7de86a55-00ea-44a6-aa0e-a9687e4fd56e","Type":"ContainerDied","Data":"67185d51fb7171ecf915a0e5cc1444cb49c59f5f6bf106aa30de576a4f77fd7f"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.928987 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" event={"ID":"7de86a55-00ea-44a6-aa0e-a9687e4fd56e","Type":"ContainerStarted","Data":"9a34d78b6f4660824d19f02e3ae0ca712d353b9658d28cab6615690b6d24eede"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.931564 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerStarted","Data":"376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42"} Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.940432 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54d58d7797-fnl7d" event={"ID":"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8","Type":"ContainerStarted","Data":"7b84e923704c763160cc6dc7dcf10d980b92a1d145d6d3efe8aa86f359d75b89"} Dec 10 19:17:02 crc kubenswrapper[4894]: E1210 19:17:02.944608 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current\\\"\"" pod="openstack/cloudkitty-db-sync-vprjr" podUID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.962990 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4gmfn" podStartSLOduration=11.202580159 podStartE2EDuration="39.962968622s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="2025-12-10 19:16:25.359854461 +0000 UTC m=+1268.154702263" lastFinishedPulling="2025-12-10 19:16:54.120242934 +0000 UTC m=+1296.915090726" observedRunningTime="2025-12-10 19:17:02.940024731 +0000 UTC m=+1305.734872523" watchObservedRunningTime="2025-12-10 19:17:02.962968622 +0000 UTC m=+1305.757816414" Dec 10 19:17:02 crc kubenswrapper[4894]: I1210 19:17:02.994660 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b2w4f" podStartSLOduration=23.994640857 podStartE2EDuration="23.994640857s" podCreationTimestamp="2025-12-10 19:16:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:02.978947878 +0000 UTC m=+1305.773795690" watchObservedRunningTime="2025-12-10 19:17:02.994640857 +0000 UTC m=+1305.789488649" Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.066655 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8659fd4574-zvfv8"] Dec 10 19:17:03 crc kubenswrapper[4894]: W1210 19:17:03.095408 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaadf98aa_0c6c_4d24_a847_e016114e615c.slice/crio-fc765c78b133ef20d11652b28d0271dce3045a031cb39c416e4fa849e641a250 WatchSource:0}: Error finding container fc765c78b133ef20d11652b28d0271dce3045a031cb39c416e4fa849e641a250: Status 404 returned error can't find the container with id fc765c78b133ef20d11652b28d0271dce3045a031cb39c416e4fa849e641a250 Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.955756 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" event={"ID":"7de86a55-00ea-44a6-aa0e-a9687e4fd56e","Type":"ContainerStarted","Data":"d5d399bc9afb03fcd324b2bce9717cc0cfb5a00d78e5c367448eb09a4d1da809"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.956504 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.960277 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9874ce74-c8e9-4719-802b-81a69271131a","Type":"ContainerStarted","Data":"a405b6ece97e13ee5a67ec751dd87c371f7281cf5e0cdd07d2c2400161b5d583"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.963671 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54d58d7797-fnl7d" event={"ID":"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8","Type":"ContainerStarted","Data":"33e1c159ae122130614a37d12925a1bcc9698dc84e6456298e4e84b965443fbe"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.963704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54d58d7797-fnl7d" event={"ID":"8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8","Type":"ContainerStarted","Data":"5bc741cd3907b129f02684b26ced3d7c522767704e498543966cf2f28f9b8c4d"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.963817 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.975599 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ca7b0207-4a62-4299-aa2a-14244c283866","Type":"ContainerStarted","Data":"ddb59b102f6178428ef1393bd729c15c7b7c711dd5fbc45982cbd4861ea55c73"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.975649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ca7b0207-4a62-4299-aa2a-14244c283866","Type":"ContainerStarted","Data":"cc88e7e330e027f3d754ca2ed27e1d86d3653e84fc0356b23537a20c616833a3"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.988936 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8659fd4574-zvfv8" event={"ID":"aadf98aa-0c6c-4d24-a847-e016114e615c","Type":"ContainerStarted","Data":"0ef8686844d0e28e9c3e28d449ff88c64f666e286d4903453a871c2aeb8353c2"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.989011 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8659fd4574-zvfv8" event={"ID":"aadf98aa-0c6c-4d24-a847-e016114e615c","Type":"ContainerStarted","Data":"dc540885ef8c2396c55fb45b6eed1e10b5a68ef93cd78c420144ae3fff163040"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.989026 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8659fd4574-zvfv8" event={"ID":"aadf98aa-0c6c-4d24-a847-e016114e615c","Type":"ContainerStarted","Data":"fc765c78b133ef20d11652b28d0271dce3045a031cb39c416e4fa849e641a250"} Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.989372 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:17:03 crc kubenswrapper[4894]: I1210 19:17:03.999356 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" podStartSLOduration=10.999331177 podStartE2EDuration="10.999331177s" podCreationTimestamp="2025-12-10 19:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:03.987761829 +0000 UTC m=+1306.782609621" watchObservedRunningTime="2025-12-10 19:17:03.999331177 +0000 UTC m=+1306.794178969" Dec 10 19:17:04 crc kubenswrapper[4894]: I1210 19:17:04.021363 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8659fd4574-zvfv8" podStartSLOduration=11.021345064 podStartE2EDuration="11.021345064s" podCreationTimestamp="2025-12-10 19:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:04.012170199 +0000 UTC m=+1306.807017991" watchObservedRunningTime="2025-12-10 19:17:04.021345064 +0000 UTC m=+1306.816192856" Dec 10 19:17:04 crc kubenswrapper[4894]: I1210 19:17:04.044359 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54d58d7797-fnl7d" podStartSLOduration=9.044335237 podStartE2EDuration="9.044335237s" podCreationTimestamp="2025-12-10 19:16:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:04.034740561 +0000 UTC m=+1306.829588363" watchObservedRunningTime="2025-12-10 19:17:04.044335237 +0000 UTC m=+1306.839183019" Dec 10 19:17:04 crc kubenswrapper[4894]: I1210 19:17:04.066204 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.06618169 podStartE2EDuration="12.06618169s" podCreationTimestamp="2025-12-10 19:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:04.057197461 +0000 UTC m=+1306.852045253" watchObservedRunningTime="2025-12-10 19:17:04.06618169 +0000 UTC m=+1306.861029482" Dec 10 19:17:04 crc kubenswrapper[4894]: I1210 19:17:04.091675 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=26.09165841 podStartE2EDuration="26.09165841s" podCreationTimestamp="2025-12-10 19:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:04.073553847 +0000 UTC m=+1306.868401649" watchObservedRunningTime="2025-12-10 19:17:04.09165841 +0000 UTC m=+1306.886506212" Dec 10 19:17:04 crc kubenswrapper[4894]: I1210 19:17:04.997720 4894 generic.go:334] "Generic (PLEG): container finished" podID="7a781080-8482-4915-9d13-b88d1785f7f3" containerID="6eeddc13ccc3804f8d502f50d581d43f84041df7a038c0e510a577216b5c622b" exitCode=0 Dec 10 19:17:04 crc kubenswrapper[4894]: I1210 19:17:04.997786 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4gmfn" event={"ID":"7a781080-8482-4915-9d13-b88d1785f7f3","Type":"ContainerDied","Data":"6eeddc13ccc3804f8d502f50d581d43f84041df7a038c0e510a577216b5c622b"} Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.432507 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4gmfn" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.565694 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-scripts\") pod \"7a781080-8482-4915-9d13-b88d1785f7f3\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.565749 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a781080-8482-4915-9d13-b88d1785f7f3-logs\") pod \"7a781080-8482-4915-9d13-b88d1785f7f3\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.565811 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-config-data\") pod \"7a781080-8482-4915-9d13-b88d1785f7f3\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.565869 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knrsg\" (UniqueName: \"kubernetes.io/projected/7a781080-8482-4915-9d13-b88d1785f7f3-kube-api-access-knrsg\") pod \"7a781080-8482-4915-9d13-b88d1785f7f3\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.566002 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle\") pod \"7a781080-8482-4915-9d13-b88d1785f7f3\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.566261 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a781080-8482-4915-9d13-b88d1785f7f3-logs" (OuterVolumeSpecName: "logs") pod "7a781080-8482-4915-9d13-b88d1785f7f3" (UID: "7a781080-8482-4915-9d13-b88d1785f7f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.566511 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a781080-8482-4915-9d13-b88d1785f7f3-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.572083 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-scripts" (OuterVolumeSpecName: "scripts") pod "7a781080-8482-4915-9d13-b88d1785f7f3" (UID: "7a781080-8482-4915-9d13-b88d1785f7f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.572128 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a781080-8482-4915-9d13-b88d1785f7f3-kube-api-access-knrsg" (OuterVolumeSpecName: "kube-api-access-knrsg") pod "7a781080-8482-4915-9d13-b88d1785f7f3" (UID: "7a781080-8482-4915-9d13-b88d1785f7f3"). InnerVolumeSpecName "kube-api-access-knrsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:06 crc kubenswrapper[4894]: E1210 19:17:06.598349 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle podName:7a781080-8482-4915-9d13-b88d1785f7f3 nodeName:}" failed. No retries permitted until 2025-12-10 19:17:07.098319615 +0000 UTC m=+1309.893167407 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle") pod "7a781080-8482-4915-9d13-b88d1785f7f3" (UID: "7a781080-8482-4915-9d13-b88d1785f7f3") : error deleting /var/lib/kubelet/pods/7a781080-8482-4915-9d13-b88d1785f7f3/volume-subpaths: remove /var/lib/kubelet/pods/7a781080-8482-4915-9d13-b88d1785f7f3/volume-subpaths: no such file or directory Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.604840 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-config-data" (OuterVolumeSpecName: "config-data") pod "7a781080-8482-4915-9d13-b88d1785f7f3" (UID: "7a781080-8482-4915-9d13-b88d1785f7f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.668972 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.669013 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:06 crc kubenswrapper[4894]: I1210 19:17:06.669025 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knrsg\" (UniqueName: \"kubernetes.io/projected/7a781080-8482-4915-9d13-b88d1785f7f3-kube-api-access-knrsg\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.071561 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4gmfn" event={"ID":"7a781080-8482-4915-9d13-b88d1785f7f3","Type":"ContainerDied","Data":"8382f482fbe41a5071517e5ecbb40b36f92fc2b058791eb194422a2fd5329539"} Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.071599 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8382f482fbe41a5071517e5ecbb40b36f92fc2b058791eb194422a2fd5329539" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.071606 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4gmfn" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.168514 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7ff967f45b-c9flb"] Dec 10 19:17:07 crc kubenswrapper[4894]: E1210 19:17:07.169074 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a781080-8482-4915-9d13-b88d1785f7f3" containerName="placement-db-sync" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.169094 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a781080-8482-4915-9d13-b88d1785f7f3" containerName="placement-db-sync" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.169295 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a781080-8482-4915-9d13-b88d1785f7f3" containerName="placement-db-sync" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.170644 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.173708 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.175278 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.178137 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle\") pod \"7a781080-8482-4915-9d13-b88d1785f7f3\" (UID: \"7a781080-8482-4915-9d13-b88d1785f7f3\") " Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.188541 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7ff967f45b-c9flb"] Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.214748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a781080-8482-4915-9d13-b88d1785f7f3" (UID: "7a781080-8482-4915-9d13-b88d1785f7f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.280237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-internal-tls-certs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.280444 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzq8r\" (UniqueName: \"kubernetes.io/projected/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-kube-api-access-xzq8r\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.280559 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-scripts\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.280600 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-combined-ca-bundle\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.280965 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-logs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.281027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-config-data\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.281074 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-public-tls-certs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.281206 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a781080-8482-4915-9d13-b88d1785f7f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.382921 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-config-data\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.382972 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-public-tls-certs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.383010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-internal-tls-certs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.383087 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzq8r\" (UniqueName: \"kubernetes.io/projected/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-kube-api-access-xzq8r\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.383129 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-scripts\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.383157 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-combined-ca-bundle\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.383212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-logs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.383645 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-logs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.387419 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-public-tls-certs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.387622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-scripts\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.388209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-internal-tls-certs\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.388345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-combined-ca-bundle\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.389793 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-config-data\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.399885 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzq8r\" (UniqueName: \"kubernetes.io/projected/6a3f07cf-68dd-4c05-9cdc-2e4dd6031070-kube-api-access-xzq8r\") pod \"placement-7ff967f45b-c9flb\" (UID: \"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070\") " pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:07 crc kubenswrapper[4894]: I1210 19:17:07.548918 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:08 crc kubenswrapper[4894]: I1210 19:17:08.064638 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7ff967f45b-c9flb"] Dec 10 19:17:08 crc kubenswrapper[4894]: W1210 19:17:08.068261 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a3f07cf_68dd_4c05_9cdc_2e4dd6031070.slice/crio-dbee08de08d8aa166be51553f69513d92e9ab6fd3627b0d57c2e45620e4c1e0b WatchSource:0}: Error finding container dbee08de08d8aa166be51553f69513d92e9ab6fd3627b0d57c2e45620e4c1e0b: Status 404 returned error can't find the container with id dbee08de08d8aa166be51553f69513d92e9ab6fd3627b0d57c2e45620e4c1e0b Dec 10 19:17:08 crc kubenswrapper[4894]: I1210 19:17:08.080055 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7ff967f45b-c9flb" event={"ID":"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070","Type":"ContainerStarted","Data":"dbee08de08d8aa166be51553f69513d92e9ab6fd3627b0d57c2e45620e4c1e0b"} Dec 10 19:17:08 crc kubenswrapper[4894]: I1210 19:17:08.727922 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:17:08 crc kubenswrapper[4894]: I1210 19:17:08.807581 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-9tcn9"] Dec 10 19:17:08 crc kubenswrapper[4894]: I1210 19:17:08.807840 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="dnsmasq-dns" containerID="cri-o://704c9a990ba89ff9bc86b6cec93f3ac4880b14bdb45ea31107ce5d32d2eaf7e7" gracePeriod=10 Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.101035 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerStarted","Data":"caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50"} Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.102745 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5765m" event={"ID":"bedfd09b-1fdb-43e9-88d7-47fd25103af6","Type":"ContainerStarted","Data":"a4109b1867a2f842962e0132f81de7089e0ec454657147bb1b03313921a51eb2"} Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.172707 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.173055 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.173065 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.173077 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.207013 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.224353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:17:09 crc kubenswrapper[4894]: I1210 19:17:09.254664 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.162:5353: connect: connection refused" Dec 10 19:17:10 crc kubenswrapper[4894]: I1210 19:17:10.113117 4894 generic.go:334] "Generic (PLEG): container finished" podID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerID="704c9a990ba89ff9bc86b6cec93f3ac4880b14bdb45ea31107ce5d32d2eaf7e7" exitCode=0 Dec 10 19:17:10 crc kubenswrapper[4894]: I1210 19:17:10.113159 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" event={"ID":"b6a2edcb-fcd4-41c4-8063-0afa95a99bda","Type":"ContainerDied","Data":"704c9a990ba89ff9bc86b6cec93f3ac4880b14bdb45ea31107ce5d32d2eaf7e7"} Dec 10 19:17:10 crc kubenswrapper[4894]: I1210 19:17:10.115136 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7ff967f45b-c9flb" event={"ID":"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070","Type":"ContainerStarted","Data":"925302f0b739f5d97956cb3bf3cba2c31fe9cca3460e7628ee543860c0042c70"} Dec 10 19:17:10 crc kubenswrapper[4894]: I1210 19:17:10.142582 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5765m" podStartSLOduration=7.814969548 podStartE2EDuration="47.142559478s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="2025-12-10 19:16:25.091664668 +0000 UTC m=+1267.886512460" lastFinishedPulling="2025-12-10 19:17:04.419254598 +0000 UTC m=+1307.214102390" observedRunningTime="2025-12-10 19:17:10.135250673 +0000 UTC m=+1312.930098485" watchObservedRunningTime="2025-12-10 19:17:10.142559478 +0000 UTC m=+1312.937407270" Dec 10 19:17:10 crc kubenswrapper[4894]: I1210 19:17:10.963783 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.060591 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-svc\") pod \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.060677 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzjgh\" (UniqueName: \"kubernetes.io/projected/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-kube-api-access-kzjgh\") pod \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.060709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-config\") pod \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.060751 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-nb\") pod \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.060775 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-swift-storage-0\") pod \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.060791 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-sb\") pod \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\" (UID: \"b6a2edcb-fcd4-41c4-8063-0afa95a99bda\") " Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.084196 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-kube-api-access-kzjgh" (OuterVolumeSpecName: "kube-api-access-kzjgh") pod "b6a2edcb-fcd4-41c4-8063-0afa95a99bda" (UID: "b6a2edcb-fcd4-41c4-8063-0afa95a99bda"). InnerVolumeSpecName "kube-api-access-kzjgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.127700 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b6a2edcb-fcd4-41c4-8063-0afa95a99bda" (UID: "b6a2edcb-fcd4-41c4-8063-0afa95a99bda"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.138300 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b6a2edcb-fcd4-41c4-8063-0afa95a99bda" (UID: "b6a2edcb-fcd4-41c4-8063-0afa95a99bda"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.141700 4894 generic.go:334] "Generic (PLEG): container finished" podID="8aac2573-2b49-4dcf-b808-d1adc6435bc9" containerID="199379249cc0934c4aaa80bdb58285e0403aecb4e7f339830f8d02331bf82029" exitCode=0 Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.141765 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b2w4f" event={"ID":"8aac2573-2b49-4dcf-b808-d1adc6435bc9","Type":"ContainerDied","Data":"199379249cc0934c4aaa80bdb58285e0403aecb4e7f339830f8d02331bf82029"} Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.147376 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" event={"ID":"b6a2edcb-fcd4-41c4-8063-0afa95a99bda","Type":"ContainerDied","Data":"0dfc3d2bd2c0ecd86660f6bad0e63e5c170d6ae70bcd1ccd5e63c61540542bab"} Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.147410 4894 scope.go:117] "RemoveContainer" containerID="704c9a990ba89ff9bc86b6cec93f3ac4880b14bdb45ea31107ce5d32d2eaf7e7" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.147499 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-9tcn9" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.165680 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzjgh\" (UniqueName: \"kubernetes.io/projected/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-kube-api-access-kzjgh\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.165732 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.165741 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.166080 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7ff967f45b-c9flb" event={"ID":"6a3f07cf-68dd-4c05-9cdc-2e4dd6031070","Type":"ContainerStarted","Data":"5ffa294e51d6fde36af04f3117e2ef5c958bb3694709919a49c42361e8b4855e"} Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.167532 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.167662 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.206205 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7ff967f45b-c9flb" podStartSLOduration=4.206185989 podStartE2EDuration="4.206185989s" podCreationTimestamp="2025-12-10 19:17:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:11.195363091 +0000 UTC m=+1313.990210893" watchObservedRunningTime="2025-12-10 19:17:11.206185989 +0000 UTC m=+1314.001033781" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.214524 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b6a2edcb-fcd4-41c4-8063-0afa95a99bda" (UID: "b6a2edcb-fcd4-41c4-8063-0afa95a99bda"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.216102 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b6a2edcb-fcd4-41c4-8063-0afa95a99bda" (UID: "b6a2edcb-fcd4-41c4-8063-0afa95a99bda"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.219775 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-config" (OuterVolumeSpecName: "config") pod "b6a2edcb-fcd4-41c4-8063-0afa95a99bda" (UID: "b6a2edcb-fcd4-41c4-8063-0afa95a99bda"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.263003 4894 scope.go:117] "RemoveContainer" containerID="0de6da4072a02827caeeb10d6384203d342b373f0a395dd2915d18044ca9806b" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.268560 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.268671 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.268741 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6a2edcb-fcd4-41c4-8063-0afa95a99bda-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.487212 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-9tcn9"] Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.493714 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-9tcn9"] Dec 10 19:17:11 crc kubenswrapper[4894]: I1210 19:17:11.569285 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" path="/var/lib/kubelet/pods/b6a2edcb-fcd4-41c4-8063-0afa95a99bda/volumes" Dec 10 19:17:12 crc kubenswrapper[4894]: I1210 19:17:12.179463 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5bq9r" event={"ID":"6524d3b3-56ee-4a03-84a5-dfa9c4c27926","Type":"ContainerStarted","Data":"5c683b8f14854e3cedf5e129dd71f810dad9575381ca6f19aecd8e7f81cc44e7"} Dec 10 19:17:12 crc kubenswrapper[4894]: I1210 19:17:12.333071 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:17:12 crc kubenswrapper[4894]: I1210 19:17:12.333365 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:17:12 crc kubenswrapper[4894]: I1210 19:17:12.369863 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-5bq9r" podStartSLOduration=4.207414725 podStartE2EDuration="49.36982601s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="2025-12-10 19:16:25.273056116 +0000 UTC m=+1268.067903908" lastFinishedPulling="2025-12-10 19:17:10.435467381 +0000 UTC m=+1313.230315193" observedRunningTime="2025-12-10 19:17:12.206186385 +0000 UTC m=+1315.001034177" watchObservedRunningTime="2025-12-10 19:17:12.36982601 +0000 UTC m=+1315.164673822" Dec 10 19:17:12 crc kubenswrapper[4894]: I1210 19:17:12.606049 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:17:13 crc kubenswrapper[4894]: I1210 19:17:13.201644 4894 generic.go:334] "Generic (PLEG): container finished" podID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" containerID="a4109b1867a2f842962e0132f81de7089e0ec454657147bb1b03313921a51eb2" exitCode=0 Dec 10 19:17:13 crc kubenswrapper[4894]: I1210 19:17:13.203498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5765m" event={"ID":"bedfd09b-1fdb-43e9-88d7-47fd25103af6","Type":"ContainerDied","Data":"a4109b1867a2f842962e0132f81de7089e0ec454657147bb1b03313921a51eb2"} Dec 10 19:17:13 crc kubenswrapper[4894]: I1210 19:17:13.400516 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:13 crc kubenswrapper[4894]: I1210 19:17:13.400565 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:13 crc kubenswrapper[4894]: I1210 19:17:13.440322 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:13 crc kubenswrapper[4894]: I1210 19:17:13.470159 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.211591 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.211875 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.544278 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.561639 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.655743 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-credential-keys\") pod \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.655895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-combined-ca-bundle\") pod \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.656027 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tx5t\" (UniqueName: \"kubernetes.io/projected/8aac2573-2b49-4dcf-b808-d1adc6435bc9-kube-api-access-4tx5t\") pod \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.656067 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-fernet-keys\") pod \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.656110 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-config-data\") pod \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.656146 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-scripts\") pod \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\" (UID: \"8aac2573-2b49-4dcf-b808-d1adc6435bc9\") " Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.669069 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aac2573-2b49-4dcf-b808-d1adc6435bc9-kube-api-access-4tx5t" (OuterVolumeSpecName: "kube-api-access-4tx5t") pod "8aac2573-2b49-4dcf-b808-d1adc6435bc9" (UID: "8aac2573-2b49-4dcf-b808-d1adc6435bc9"). InnerVolumeSpecName "kube-api-access-4tx5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.680503 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8aac2573-2b49-4dcf-b808-d1adc6435bc9" (UID: "8aac2573-2b49-4dcf-b808-d1adc6435bc9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.686326 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8aac2573-2b49-4dcf-b808-d1adc6435bc9" (UID: "8aac2573-2b49-4dcf-b808-d1adc6435bc9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.695022 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-scripts" (OuterVolumeSpecName: "scripts") pod "8aac2573-2b49-4dcf-b808-d1adc6435bc9" (UID: "8aac2573-2b49-4dcf-b808-d1adc6435bc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.724521 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8aac2573-2b49-4dcf-b808-d1adc6435bc9" (UID: "8aac2573-2b49-4dcf-b808-d1adc6435bc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.760446 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.760489 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tx5t\" (UniqueName: \"kubernetes.io/projected/8aac2573-2b49-4dcf-b808-d1adc6435bc9-kube-api-access-4tx5t\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.760507 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.760518 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.760533 4894 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.790948 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-config-data" (OuterVolumeSpecName: "config-data") pod "8aac2573-2b49-4dcf-b808-d1adc6435bc9" (UID: "8aac2573-2b49-4dcf-b808-d1adc6435bc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:14 crc kubenswrapper[4894]: I1210 19:17:14.863191 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aac2573-2b49-4dcf-b808-d1adc6435bc9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.223985 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b2w4f" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.224239 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b2w4f" event={"ID":"8aac2573-2b49-4dcf-b808-d1adc6435bc9","Type":"ContainerDied","Data":"168dd2a6938a264a69ae1d2ffdf9c1546b577b14f80d5e331c0260aa9b905d09"} Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.225868 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="168dd2a6938a264a69ae1d2ffdf9c1546b577b14f80d5e331c0260aa9b905d09" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.672124 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6bf6fc85f8-h5xlr"] Dec 10 19:17:15 crc kubenswrapper[4894]: E1210 19:17:15.672686 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="dnsmasq-dns" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.672775 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="dnsmasq-dns" Dec 10 19:17:15 crc kubenswrapper[4894]: E1210 19:17:15.672858 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="init" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.672910 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="init" Dec 10 19:17:15 crc kubenswrapper[4894]: E1210 19:17:15.672977 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aac2573-2b49-4dcf-b808-d1adc6435bc9" containerName="keystone-bootstrap" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.673024 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aac2573-2b49-4dcf-b808-d1adc6435bc9" containerName="keystone-bootstrap" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.673233 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aac2573-2b49-4dcf-b808-d1adc6435bc9" containerName="keystone-bootstrap" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.673298 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6a2edcb-fcd4-41c4-8063-0afa95a99bda" containerName="dnsmasq-dns" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.673985 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.676875 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ph6mx" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.677031 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.677030 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.677247 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.677627 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.679147 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.720075 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bf6fc85f8-h5xlr"] Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779504 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-fernet-keys\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779556 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-internal-tls-certs\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779585 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-public-tls-certs\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779613 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-scripts\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779629 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-config-data\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779664 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m58fv\" (UniqueName: \"kubernetes.io/projected/06f64e80-03ab-47da-8dd6-670aaa6f1a32-kube-api-access-m58fv\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779702 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-combined-ca-bundle\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.779748 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-credential-keys\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-fernet-keys\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881756 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-internal-tls-certs\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-public-tls-certs\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881813 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-scripts\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881829 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-config-data\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881893 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m58fv\" (UniqueName: \"kubernetes.io/projected/06f64e80-03ab-47da-8dd6-670aaa6f1a32-kube-api-access-m58fv\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881932 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-combined-ca-bundle\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.881982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-credential-keys\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.887130 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-fernet-keys\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.887515 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-scripts\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.887773 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-credential-keys\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.891238 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-public-tls-certs\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.891498 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-config-data\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.891602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-combined-ca-bundle\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.891982 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06f64e80-03ab-47da-8dd6-670aaa6f1a32-internal-tls-certs\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.906925 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m58fv\" (UniqueName: \"kubernetes.io/projected/06f64e80-03ab-47da-8dd6-670aaa6f1a32-kube-api-access-m58fv\") pod \"keystone-6bf6fc85f8-h5xlr\" (UID: \"06f64e80-03ab-47da-8dd6-670aaa6f1a32\") " pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:15 crc kubenswrapper[4894]: I1210 19:17:15.993303 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.150312 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.150669 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.156094 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.252624 4894 generic.go:334] "Generic (PLEG): container finished" podID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" containerID="5c683b8f14854e3cedf5e129dd71f810dad9575381ca6f19aecd8e7f81cc44e7" exitCode=0 Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.253989 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5bq9r" event={"ID":"6524d3b3-56ee-4a03-84a5-dfa9c4c27926","Type":"ContainerDied","Data":"5c683b8f14854e3cedf5e129dd71f810dad9575381ca6f19aecd8e7f81cc44e7"} Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.628535 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5765m" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.735962 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-db-sync-config-data\") pod \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.736038 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-combined-ca-bundle\") pod \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.736075 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdvk8\" (UniqueName: \"kubernetes.io/projected/bedfd09b-1fdb-43e9-88d7-47fd25103af6-kube-api-access-hdvk8\") pod \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\" (UID: \"bedfd09b-1fdb-43e9-88d7-47fd25103af6\") " Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.742800 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bedfd09b-1fdb-43e9-88d7-47fd25103af6-kube-api-access-hdvk8" (OuterVolumeSpecName: "kube-api-access-hdvk8") pod "bedfd09b-1fdb-43e9-88d7-47fd25103af6" (UID: "bedfd09b-1fdb-43e9-88d7-47fd25103af6"). InnerVolumeSpecName "kube-api-access-hdvk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.759638 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bedfd09b-1fdb-43e9-88d7-47fd25103af6" (UID: "bedfd09b-1fdb-43e9-88d7-47fd25103af6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.768553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bedfd09b-1fdb-43e9-88d7-47fd25103af6" (UID: "bedfd09b-1fdb-43e9-88d7-47fd25103af6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.838942 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.838975 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bedfd09b-1fdb-43e9-88d7-47fd25103af6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:17 crc kubenswrapper[4894]: I1210 19:17:17.838988 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdvk8\" (UniqueName: \"kubernetes.io/projected/bedfd09b-1fdb-43e9-88d7-47fd25103af6-kube-api-access-hdvk8\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.011031 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bf6fc85f8-h5xlr"] Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.269958 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bf6fc85f8-h5xlr" event={"ID":"06f64e80-03ab-47da-8dd6-670aaa6f1a32","Type":"ContainerStarted","Data":"36a5a12d4cefa1a7052baf03ec46669302b0c1c51d4b8dab40f77e79be0442fa"} Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.270001 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bf6fc85f8-h5xlr" event={"ID":"06f64e80-03ab-47da-8dd6-670aaa6f1a32","Type":"ContainerStarted","Data":"d5fc406e99074b25c11846aa09a60fb7174393ae7dacd8afb9406b1cce8e5319"} Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.270578 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.273484 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5765m" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.273691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5765m" event={"ID":"bedfd09b-1fdb-43e9-88d7-47fd25103af6","Type":"ContainerDied","Data":"1b1057939bc0c4515ff2c77aced64b39de3ccc0fbeddd3ff96adc6d6c7456a53"} Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.273723 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b1057939bc0c4515ff2c77aced64b39de3ccc0fbeddd3ff96adc6d6c7456a53" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.315345 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6bf6fc85f8-h5xlr" podStartSLOduration=3.315320796 podStartE2EDuration="3.315320796s" podCreationTimestamp="2025-12-10 19:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:18.300560631 +0000 UTC m=+1321.095408433" watchObservedRunningTime="2025-12-10 19:17:18.315320796 +0000 UTC m=+1321.110168588" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.727762 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.773164 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-combined-ca-bundle\") pod \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.774315 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-etc-machine-id\") pod \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.774399 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-db-sync-config-data\") pod \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.774533 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-config-data\") pod \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.774622 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcdch\" (UniqueName: \"kubernetes.io/projected/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-kube-api-access-gcdch\") pod \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.774698 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-scripts\") pod \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\" (UID: \"6524d3b3-56ee-4a03-84a5-dfa9c4c27926\") " Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.774413 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6524d3b3-56ee-4a03-84a5-dfa9c4c27926" (UID: "6524d3b3-56ee-4a03-84a5-dfa9c4c27926"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.776738 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.784310 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-kube-api-access-gcdch" (OuterVolumeSpecName: "kube-api-access-gcdch") pod "6524d3b3-56ee-4a03-84a5-dfa9c4c27926" (UID: "6524d3b3-56ee-4a03-84a5-dfa9c4c27926"). InnerVolumeSpecName "kube-api-access-gcdch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.784658 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-scripts" (OuterVolumeSpecName: "scripts") pod "6524d3b3-56ee-4a03-84a5-dfa9c4c27926" (UID: "6524d3b3-56ee-4a03-84a5-dfa9c4c27926"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.792920 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6524d3b3-56ee-4a03-84a5-dfa9c4c27926" (UID: "6524d3b3-56ee-4a03-84a5-dfa9c4c27926"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.878371 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.878404 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcdch\" (UniqueName: \"kubernetes.io/projected/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-kube-api-access-gcdch\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.878413 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.935054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6524d3b3-56ee-4a03-84a5-dfa9c4c27926" (UID: "6524d3b3-56ee-4a03-84a5-dfa9c4c27926"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.937981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-config-data" (OuterVolumeSpecName: "config-data") pod "6524d3b3-56ee-4a03-84a5-dfa9c4c27926" (UID: "6524d3b3-56ee-4a03-84a5-dfa9c4c27926"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.980658 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:18 crc kubenswrapper[4894]: I1210 19:17:18.980689 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524d3b3-56ee-4a03-84a5-dfa9c4c27926-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.046565 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f9794c7b8-gr6xz"] Dec 10 19:17:19 crc kubenswrapper[4894]: E1210 19:17:19.047075 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" containerName="barbican-db-sync" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.047106 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" containerName="barbican-db-sync" Dec 10 19:17:19 crc kubenswrapper[4894]: E1210 19:17:19.047117 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" containerName="cinder-db-sync" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.047123 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" containerName="cinder-db-sync" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.047361 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" containerName="barbican-db-sync" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.047385 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" containerName="cinder-db-sync" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.048631 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.065564 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7cbb59f5d9-2cqct"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.068678 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.074325 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.074526 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rt8vq" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.074655 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.074758 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.083521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5t6x\" (UniqueName: \"kubernetes.io/projected/30ab0905-d039-4f32-9740-3b7bffaab6ce-kube-api-access-f5t6x\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.083597 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-config-data\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.083702 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ab0905-d039-4f32-9740-3b7bffaab6ce-logs\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.083722 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-combined-ca-bundle\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.083784 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-config-data-custom\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.085591 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7cbb59f5d9-2cqct"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.097378 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f9794c7b8-gr6xz"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.185979 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-config-data\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186050 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-combined-ca-bundle\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186114 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-config-data\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ab0905-d039-4f32-9740-3b7bffaab6ce-logs\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186223 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-combined-ca-bundle\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186271 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-config-data-custom\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186324 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-logs\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186351 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54wsb\" (UniqueName: \"kubernetes.io/projected/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-kube-api-access-54wsb\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186381 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5t6x\" (UniqueName: \"kubernetes.io/projected/30ab0905-d039-4f32-9740-3b7bffaab6ce-kube-api-access-f5t6x\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.186404 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-config-data-custom\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.189528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ab0905-d039-4f32-9740-3b7bffaab6ce-logs\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.193494 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-config-data\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.202356 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-config-data-custom\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.203467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab0905-d039-4f32-9740-3b7bffaab6ce-combined-ca-bundle\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.251565 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5t6x\" (UniqueName: \"kubernetes.io/projected/30ab0905-d039-4f32-9740-3b7bffaab6ce-kube-api-access-f5t6x\") pod \"barbican-keystone-listener-7f9794c7b8-gr6xz\" (UID: \"30ab0905-d039-4f32-9740-3b7bffaab6ce\") " pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.281110 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-fc5kz"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.294735 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.294762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-logs\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.294826 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54wsb\" (UniqueName: \"kubernetes.io/projected/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-kube-api-access-54wsb\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.294920 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-config-data-custom\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.295053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-combined-ca-bundle\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.295131 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-config-data\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.295257 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-logs\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.306464 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-config-data-custom\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.316318 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-config-data\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.333547 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-combined-ca-bundle\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.342571 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54wsb\" (UniqueName: \"kubernetes.io/projected/25d1c4d9-024f-4f80-97e8-8b3c7fb726b0-kube-api-access-54wsb\") pod \"barbican-worker-7cbb59f5d9-2cqct\" (UID: \"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0\") " pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.343367 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vprjr" event={"ID":"bc3ed2c8-9e2c-443b-bc22-a39397986e13","Type":"ContainerStarted","Data":"428fe02775bef18d060f3c3ecbbd4cbb20b6c059bfa694e2f62243ed2fbb2e90"} Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.347889 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-fc5kz"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.372981 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.375168 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5bq9r" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.375211 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5bq9r" event={"ID":"6524d3b3-56ee-4a03-84a5-dfa9c4c27926","Type":"ContainerDied","Data":"7cae81252524e5eacb5c1bf72de9e6a5ba083cf357ce875e8503b651d7205bda"} Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.375242 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cae81252524e5eacb5c1bf72de9e6a5ba083cf357ce875e8503b651d7205bda" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.393377 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7cbb59f5d9-2cqct" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.401078 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-svc\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.401258 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.401399 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.401544 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.401655 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brcq2\" (UniqueName: \"kubernetes.io/projected/77dce491-11c8-410e-80ed-7dc6a69cb925-kube-api-access-brcq2\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.401724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-config\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.410123 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerStarted","Data":"f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95"} Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.429905 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5cb6f668cd-tg9dt"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.431537 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.435584 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.448356 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cb6f668cd-tg9dt"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.464918 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-vprjr" podStartSLOduration=3.9741800229999997 podStartE2EDuration="56.464902391s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="2025-12-10 19:16:25.541860606 +0000 UTC m=+1268.336708398" lastFinishedPulling="2025-12-10 19:17:18.032582974 +0000 UTC m=+1320.827430766" observedRunningTime="2025-12-10 19:17:19.435324892 +0000 UTC m=+1322.230172684" watchObservedRunningTime="2025-12-10 19:17:19.464902391 +0000 UTC m=+1322.259750173" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509465 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509518 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509544 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-logs\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data-custom\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509649 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509666 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509709 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4q5k\" (UniqueName: \"kubernetes.io/projected/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-kube-api-access-g4q5k\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509730 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brcq2\" (UniqueName: \"kubernetes.io/projected/77dce491-11c8-410e-80ed-7dc6a69cb925-kube-api-access-brcq2\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509747 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-combined-ca-bundle\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509769 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-config\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.509811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-svc\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.510727 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.511001 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.512800 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-svc\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.512868 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-config\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.514892 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.543414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brcq2\" (UniqueName: \"kubernetes.io/projected/77dce491-11c8-410e-80ed-7dc6a69cb925-kube-api-access-brcq2\") pod \"dnsmasq-dns-688c87cc99-fc5kz\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.611672 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-logs\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.611793 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data-custom\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.611814 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.611950 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4q5k\" (UniqueName: \"kubernetes.io/projected/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-kube-api-access-g4q5k\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.611974 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-combined-ca-bundle\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.613257 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-logs\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.628898 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.631972 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-combined-ca-bundle\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.639495 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4q5k\" (UniqueName: \"kubernetes.io/projected/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-kube-api-access-g4q5k\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.651594 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data-custom\") pod \"barbican-api-5cb6f668cd-tg9dt\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.675301 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.688415 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.689992 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.693618 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ktrjs" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.693797 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.693933 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.694037 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.698693 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.753305 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.790938 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-fc5kz"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819401 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2frqm"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819546 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819587 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4tp5\" (UniqueName: \"kubernetes.io/projected/9b5f48b8-38ba-4884-b65a-46decaca2122-kube-api-access-t4tp5\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819640 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-scripts\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819658 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819687 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b5f48b8-38ba-4884-b65a-46decaca2122-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.819762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.821059 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.887910 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2frqm"] Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.921932 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.921993 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25x9r\" (UniqueName: \"kubernetes.io/projected/9a08e52d-bd79-4511-ab2e-86dfb4a42449-kube-api-access-25x9r\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922039 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922101 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922117 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-config\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922148 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4tp5\" (UniqueName: \"kubernetes.io/projected/9b5f48b8-38ba-4884-b65a-46decaca2122-kube-api-access-t4tp5\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922189 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922221 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-scripts\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922238 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922274 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b5f48b8-38ba-4884-b65a-46decaca2122-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.922385 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b5f48b8-38ba-4884-b65a-46decaca2122-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.934385 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-scripts\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.934642 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.935127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.936787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:19 crc kubenswrapper[4894]: I1210 19:17:19.947649 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4tp5\" (UniqueName: \"kubernetes.io/projected/9b5f48b8-38ba-4884-b65a-46decaca2122-kube-api-access-t4tp5\") pod \"cinder-scheduler-0\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.017517 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.024473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.024546 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.024570 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-config\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.024620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.024756 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25x9r\" (UniqueName: \"kubernetes.io/projected/9a08e52d-bd79-4511-ab2e-86dfb4a42449-kube-api-access-25x9r\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.024777 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.025598 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.026211 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.027029 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-config\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.027473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.028152 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.031562 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.033810 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.045516 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.049709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25x9r\" (UniqueName: \"kubernetes.io/projected/9a08e52d-bd79-4511-ab2e-86dfb4a42449-kube-api-access-25x9r\") pod \"dnsmasq-dns-6bb4fc677f-2frqm\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.085324 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.126891 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1c77e8f-d949-45af-9cbb-725cfe347782-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.126977 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.127026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p245\" (UniqueName: \"kubernetes.io/projected/c1c77e8f-d949-45af-9cbb-725cfe347782-kube-api-access-2p245\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.127058 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c77e8f-d949-45af-9cbb-725cfe347782-logs\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.127140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data-custom\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.127175 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-scripts\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.127220 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.185481 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.228830 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-scripts\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.228963 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.229043 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1c77e8f-d949-45af-9cbb-725cfe347782-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.229085 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.229121 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p245\" (UniqueName: \"kubernetes.io/projected/c1c77e8f-d949-45af-9cbb-725cfe347782-kube-api-access-2p245\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.229158 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c77e8f-d949-45af-9cbb-725cfe347782-logs\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.229216 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data-custom\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.229401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1c77e8f-d949-45af-9cbb-725cfe347782-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.233502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c77e8f-d949-45af-9cbb-725cfe347782-logs\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.234637 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-scripts\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.234878 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data-custom\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.236957 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.244240 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.257369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p245\" (UniqueName: \"kubernetes.io/projected/c1c77e8f-d949-45af-9cbb-725cfe347782-kube-api-access-2p245\") pod \"cinder-api-0\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.270735 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f9794c7b8-gr6xz"] Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.377895 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.393023 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7cbb59f5d9-2cqct"] Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.448644 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" event={"ID":"30ab0905-d039-4f32-9740-3b7bffaab6ce","Type":"ContainerStarted","Data":"ae9be735ec162f521bd3d8da3258d895717c545caf441236e14568771453d5c5"} Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.521422 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5cb6f668cd-tg9dt"] Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.555106 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-fc5kz"] Dec 10 19:17:20 crc kubenswrapper[4894]: W1210 19:17:20.608694 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77dce491_11c8_410e_80ed_7dc6a69cb925.slice/crio-e941edb3c07da6a8bbba0efa45967563fbc46ef6a4fe56803b97cf0b40544990 WatchSource:0}: Error finding container e941edb3c07da6a8bbba0efa45967563fbc46ef6a4fe56803b97cf0b40544990: Status 404 returned error can't find the container with id e941edb3c07da6a8bbba0efa45967563fbc46ef6a4fe56803b97cf0b40544990 Dec 10 19:17:20 crc kubenswrapper[4894]: I1210 19:17:20.856955 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:20 crc kubenswrapper[4894]: W1210 19:17:20.870806 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b5f48b8_38ba_4884_b65a_46decaca2122.slice/crio-6c1a9ac7b1fd4859c3fe52ac131b451f33e29a8c128dafd87bdab517b955a720 WatchSource:0}: Error finding container 6c1a9ac7b1fd4859c3fe52ac131b451f33e29a8c128dafd87bdab517b955a720: Status 404 returned error can't find the container with id 6c1a9ac7b1fd4859c3fe52ac131b451f33e29a8c128dafd87bdab517b955a720 Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.028248 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2frqm"] Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.046400 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.522888 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb6f668cd-tg9dt" event={"ID":"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4","Type":"ContainerStarted","Data":"ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.523228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb6f668cd-tg9dt" event={"ID":"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4","Type":"ContainerStarted","Data":"95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.523245 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb6f668cd-tg9dt" event={"ID":"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4","Type":"ContainerStarted","Data":"408c963617aa4c7876f76d0c81605ca21274548c9b7c9513b8512ba0b0e57d0e"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.523423 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.523448 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.525281 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cbb59f5d9-2cqct" event={"ID":"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0","Type":"ContainerStarted","Data":"d7c7827fe9273d93581d27f3cb586707f2bbcf3ca018cb1df1e22ce58d872258"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.533164 4894 generic.go:334] "Generic (PLEG): container finished" podID="77dce491-11c8-410e-80ed-7dc6a69cb925" containerID="af3f68eb8618f4ece6875da50a527153cd88364b7ea5a1d342beaa51c7501080" exitCode=0 Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.533215 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" event={"ID":"77dce491-11c8-410e-80ed-7dc6a69cb925","Type":"ContainerDied","Data":"af3f68eb8618f4ece6875da50a527153cd88364b7ea5a1d342beaa51c7501080"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.533235 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" event={"ID":"77dce491-11c8-410e-80ed-7dc6a69cb925","Type":"ContainerStarted","Data":"e941edb3c07da6a8bbba0efa45967563fbc46ef6a4fe56803b97cf0b40544990"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.560027 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1c77e8f-d949-45af-9cbb-725cfe347782","Type":"ContainerStarted","Data":"c6aab3d4838cbb48d8b99f8d4615b1b811d8e19db55c07d05037cbb3d6ad8a1b"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.564261 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5cb6f668cd-tg9dt" podStartSLOduration=2.56424298 podStartE2EDuration="2.56424298s" podCreationTimestamp="2025-12-10 19:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:21.546482847 +0000 UTC m=+1324.341330629" watchObservedRunningTime="2025-12-10 19:17:21.56424298 +0000 UTC m=+1324.359090772" Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.576887 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" event={"ID":"9a08e52d-bd79-4511-ab2e-86dfb4a42449","Type":"ContainerStarted","Data":"af0b44411d47272a828aaaf41fa6cec3f568e876d28b0ada56e5aa9176578d1c"} Dec 10 19:17:21 crc kubenswrapper[4894]: I1210 19:17:21.580828 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9b5f48b8-38ba-4884-b65a-46decaca2122","Type":"ContainerStarted","Data":"6c1a9ac7b1fd4859c3fe52ac131b451f33e29a8c128dafd87bdab517b955a720"} Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.309220 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.408286 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-sb\") pod \"77dce491-11c8-410e-80ed-7dc6a69cb925\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.408345 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-svc\") pod \"77dce491-11c8-410e-80ed-7dc6a69cb925\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.408364 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-config\") pod \"77dce491-11c8-410e-80ed-7dc6a69cb925\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.408388 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-swift-storage-0\") pod \"77dce491-11c8-410e-80ed-7dc6a69cb925\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.408442 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-nb\") pod \"77dce491-11c8-410e-80ed-7dc6a69cb925\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.408490 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brcq2\" (UniqueName: \"kubernetes.io/projected/77dce491-11c8-410e-80ed-7dc6a69cb925-kube-api-access-brcq2\") pod \"77dce491-11c8-410e-80ed-7dc6a69cb925\" (UID: \"77dce491-11c8-410e-80ed-7dc6a69cb925\") " Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.415212 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77dce491-11c8-410e-80ed-7dc6a69cb925-kube-api-access-brcq2" (OuterVolumeSpecName: "kube-api-access-brcq2") pod "77dce491-11c8-410e-80ed-7dc6a69cb925" (UID: "77dce491-11c8-410e-80ed-7dc6a69cb925"). InnerVolumeSpecName "kube-api-access-brcq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.444722 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "77dce491-11c8-410e-80ed-7dc6a69cb925" (UID: "77dce491-11c8-410e-80ed-7dc6a69cb925"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.448068 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "77dce491-11c8-410e-80ed-7dc6a69cb925" (UID: "77dce491-11c8-410e-80ed-7dc6a69cb925"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.450943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "77dce491-11c8-410e-80ed-7dc6a69cb925" (UID: "77dce491-11c8-410e-80ed-7dc6a69cb925"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.464506 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-config" (OuterVolumeSpecName: "config") pod "77dce491-11c8-410e-80ed-7dc6a69cb925" (UID: "77dce491-11c8-410e-80ed-7dc6a69cb925"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.466391 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "77dce491-11c8-410e-80ed-7dc6a69cb925" (UID: "77dce491-11c8-410e-80ed-7dc6a69cb925"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.510743 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brcq2\" (UniqueName: \"kubernetes.io/projected/77dce491-11c8-410e-80ed-7dc6a69cb925-kube-api-access-brcq2\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.510775 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.510785 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.510797 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.510807 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.510820 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77dce491-11c8-410e-80ed-7dc6a69cb925-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.598149 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1c77e8f-d949-45af-9cbb-725cfe347782","Type":"ContainerStarted","Data":"1a554d9c70f1b174e307856c4dda22fd9f0a701829c79a5824315b02ed5d337a"} Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.600974 4894 generic.go:334] "Generic (PLEG): container finished" podID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerID="b45df4516f13da7a744ce9eca1f46dfd5286f2bf0ca611127475e99186420f8c" exitCode=0 Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.601041 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" event={"ID":"9a08e52d-bd79-4511-ab2e-86dfb4a42449","Type":"ContainerDied","Data":"b45df4516f13da7a744ce9eca1f46dfd5286f2bf0ca611127475e99186420f8c"} Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.606180 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.610710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-fc5kz" event={"ID":"77dce491-11c8-410e-80ed-7dc6a69cb925","Type":"ContainerDied","Data":"e941edb3c07da6a8bbba0efa45967563fbc46ef6a4fe56803b97cf0b40544990"} Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.610812 4894 scope.go:117] "RemoveContainer" containerID="af3f68eb8618f4ece6875da50a527153cd88364b7ea5a1d342beaa51c7501080" Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.684085 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-fc5kz"] Dec 10 19:17:22 crc kubenswrapper[4894]: I1210 19:17:22.701737 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-fc5kz"] Dec 10 19:17:23 crc kubenswrapper[4894]: I1210 19:17:23.152279 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:23 crc kubenswrapper[4894]: I1210 19:17:23.613557 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77dce491-11c8-410e-80ed-7dc6a69cb925" path="/var/lib/kubelet/pods/77dce491-11c8-410e-80ed-7dc6a69cb925/volumes" Dec 10 19:17:23 crc kubenswrapper[4894]: I1210 19:17:23.708156 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9b5f48b8-38ba-4884-b65a-46decaca2122","Type":"ContainerStarted","Data":"73056b57e61fc801fe826cab24b6bd922c3a3dc33afb22e4a6f4d692fedfb6b6"} Dec 10 19:17:23 crc kubenswrapper[4894]: I1210 19:17:23.852245 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:17:24 crc kubenswrapper[4894]: I1210 19:17:24.740599 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" event={"ID":"30ab0905-d039-4f32-9740-3b7bffaab6ce","Type":"ContainerStarted","Data":"4c6c8a9305d1c0f634d774559eaaa1a076a58090a2b12ff703b0c5e23bb12539"} Dec 10 19:17:24 crc kubenswrapper[4894]: I1210 19:17:24.743118 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" event={"ID":"9a08e52d-bd79-4511-ab2e-86dfb4a42449","Type":"ContainerStarted","Data":"42af34c411f9b80c3afb60391e4a65aa54154b78cc246261bf626674ae222948"} Dec 10 19:17:24 crc kubenswrapper[4894]: I1210 19:17:24.743270 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:24 crc kubenswrapper[4894]: I1210 19:17:24.744527 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cbb59f5d9-2cqct" event={"ID":"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0","Type":"ContainerStarted","Data":"98bb39f0357406008f2c2ef7fc2b4e7cb5dab1b18470ad8b4fd0b270dcfa9305"} Dec 10 19:17:24 crc kubenswrapper[4894]: I1210 19:17:24.763292 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" podStartSLOduration=5.7632741240000005 podStartE2EDuration="5.763274124s" podCreationTimestamp="2025-12-10 19:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:24.758053815 +0000 UTC m=+1327.552901607" watchObservedRunningTime="2025-12-10 19:17:24.763274124 +0000 UTC m=+1327.558121916" Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.772625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cbb59f5d9-2cqct" event={"ID":"25d1c4d9-024f-4f80-97e8-8b3c7fb726b0","Type":"ContainerStarted","Data":"32be2741a95752a998bdc03b958293f8baf41a1c11c622b304fced25543f61cc"} Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.787938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" event={"ID":"30ab0905-d039-4f32-9740-3b7bffaab6ce","Type":"ContainerStarted","Data":"263788edce1362cce2e2c5372ad6eb1da215e270211ceddb33e8e098aa81bc5b"} Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.816142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1c77e8f-d949-45af-9cbb-725cfe347782","Type":"ContainerStarted","Data":"7afe097e6035d2a0dcbe9d6afb61d942ac43ffb9f7921b9e38583340c9b68afa"} Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.816525 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api-log" containerID="cri-o://1a554d9c70f1b174e307856c4dda22fd9f0a701829c79a5824315b02ed5d337a" gracePeriod=30 Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.816744 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.816775 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api" containerID="cri-o://7afe097e6035d2a0dcbe9d6afb61d942ac43ffb9f7921b9e38583340c9b68afa" gracePeriod=30 Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.818832 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f9794c7b8-gr6xz" podStartSLOduration=4.624920904 podStartE2EDuration="7.818820211s" podCreationTimestamp="2025-12-10 19:17:18 +0000 UTC" firstStartedPulling="2025-12-10 19:17:20.28139304 +0000 UTC m=+1323.076240832" lastFinishedPulling="2025-12-10 19:17:23.475292347 +0000 UTC m=+1326.270140139" observedRunningTime="2025-12-10 19:17:25.818361439 +0000 UTC m=+1328.613209241" watchObservedRunningTime="2025-12-10 19:17:25.818820211 +0000 UTC m=+1328.613668003" Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.824152 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7cbb59f5d9-2cqct" podStartSLOduration=4.807230967 podStartE2EDuration="7.824131613s" podCreationTimestamp="2025-12-10 19:17:18 +0000 UTC" firstStartedPulling="2025-12-10 19:17:20.434222127 +0000 UTC m=+1323.229069919" lastFinishedPulling="2025-12-10 19:17:23.451122773 +0000 UTC m=+1326.245970565" observedRunningTime="2025-12-10 19:17:25.803096152 +0000 UTC m=+1328.597943944" watchObservedRunningTime="2025-12-10 19:17:25.824131613 +0000 UTC m=+1328.618979405" Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.840275 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.840262403 podStartE2EDuration="6.840262403s" podCreationTimestamp="2025-12-10 19:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:25.838906566 +0000 UTC m=+1328.633754368" watchObservedRunningTime="2025-12-10 19:17:25.840262403 +0000 UTC m=+1328.635110195" Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.849161 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9b5f48b8-38ba-4884-b65a-46decaca2122","Type":"ContainerStarted","Data":"8451a2a62025ae10893821b5dd1222630c84b53b7f4687d3ca2aeddd92b3e872"} Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.872532 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54d58d7797-fnl7d" Dec 10 19:17:25 crc kubenswrapper[4894]: I1210 19:17:25.889272 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.850038757 podStartE2EDuration="6.889251849s" podCreationTimestamp="2025-12-10 19:17:19 +0000 UTC" firstStartedPulling="2025-12-10 19:17:20.872729354 +0000 UTC m=+1323.667577146" lastFinishedPulling="2025-12-10 19:17:21.911942446 +0000 UTC m=+1324.706790238" observedRunningTime="2025-12-10 19:17:25.874361443 +0000 UTC m=+1328.669209235" watchObservedRunningTime="2025-12-10 19:17:25.889251849 +0000 UTC m=+1328.684099641" Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.020896 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8659fd4574-zvfv8"] Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.021102 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8659fd4574-zvfv8" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-api" containerID="cri-o://dc540885ef8c2396c55fb45b6eed1e10b5a68ef93cd78c420144ae3fff163040" gracePeriod=30 Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.022751 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8659fd4574-zvfv8" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-httpd" containerID="cri-o://0ef8686844d0e28e9c3e28d449ff88c64f666e286d4903453a871c2aeb8353c2" gracePeriod=30 Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.864188 4894 generic.go:334] "Generic (PLEG): container finished" podID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerID="0ef8686844d0e28e9c3e28d449ff88c64f666e286d4903453a871c2aeb8353c2" exitCode=0 Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.864237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8659fd4574-zvfv8" event={"ID":"aadf98aa-0c6c-4d24-a847-e016114e615c","Type":"ContainerDied","Data":"0ef8686844d0e28e9c3e28d449ff88c64f666e286d4903453a871c2aeb8353c2"} Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.866331 4894 generic.go:334] "Generic (PLEG): container finished" podID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" containerID="428fe02775bef18d060f3c3ecbbd4cbb20b6c059bfa694e2f62243ed2fbb2e90" exitCode=0 Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.866416 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vprjr" event={"ID":"bc3ed2c8-9e2c-443b-bc22-a39397986e13","Type":"ContainerDied","Data":"428fe02775bef18d060f3c3ecbbd4cbb20b6c059bfa694e2f62243ed2fbb2e90"} Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.868005 4894 generic.go:334] "Generic (PLEG): container finished" podID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerID="7afe097e6035d2a0dcbe9d6afb61d942ac43ffb9f7921b9e38583340c9b68afa" exitCode=0 Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.868027 4894 generic.go:334] "Generic (PLEG): container finished" podID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerID="1a554d9c70f1b174e307856c4dda22fd9f0a701829c79a5824315b02ed5d337a" exitCode=143 Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.868122 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1c77e8f-d949-45af-9cbb-725cfe347782","Type":"ContainerDied","Data":"7afe097e6035d2a0dcbe9d6afb61d942ac43ffb9f7921b9e38583340c9b68afa"} Dec 10 19:17:26 crc kubenswrapper[4894]: I1210 19:17:26.868157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1c77e8f-d949-45af-9cbb-725cfe347782","Type":"ContainerDied","Data":"1a554d9c70f1b174e307856c4dda22fd9f0a701829c79a5824315b02ed5d337a"} Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.218601 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54c578ccd6-gksvt"] Dec 10 19:17:27 crc kubenswrapper[4894]: E1210 19:17:27.219026 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77dce491-11c8-410e-80ed-7dc6a69cb925" containerName="init" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.219043 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="77dce491-11c8-410e-80ed-7dc6a69cb925" containerName="init" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.219262 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="77dce491-11c8-410e-80ed-7dc6a69cb925" containerName="init" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.220391 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.224615 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.224743 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.242315 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54c578ccd6-gksvt"] Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.246727 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585f9d29-8845-45a9-9dd3-8fba7042a4e3-logs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.246863 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-config-data-custom\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.246937 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4887q\" (UniqueName: \"kubernetes.io/projected/585f9d29-8845-45a9-9dd3-8fba7042a4e3-kube-api-access-4887q\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.246996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-public-tls-certs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.247057 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-combined-ca-bundle\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.247108 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-config-data\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.247182 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-internal-tls-certs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-combined-ca-bundle\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-config-data\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-internal-tls-certs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585f9d29-8845-45a9-9dd3-8fba7042a4e3-logs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351613 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-config-data-custom\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351667 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4887q\" (UniqueName: \"kubernetes.io/projected/585f9d29-8845-45a9-9dd3-8fba7042a4e3-kube-api-access-4887q\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.351704 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-public-tls-certs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.352420 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585f9d29-8845-45a9-9dd3-8fba7042a4e3-logs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.357918 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-public-tls-certs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.358288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-config-data-custom\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.358487 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-internal-tls-certs\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.373237 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-combined-ca-bundle\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.374380 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585f9d29-8845-45a9-9dd3-8fba7042a4e3-config-data\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.385138 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4887q\" (UniqueName: \"kubernetes.io/projected/585f9d29-8845-45a9-9dd3-8fba7042a4e3-kube-api-access-4887q\") pod \"barbican-api-54c578ccd6-gksvt\" (UID: \"585f9d29-8845-45a9-9dd3-8fba7042a4e3\") " pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:27 crc kubenswrapper[4894]: I1210 19:17:27.567514 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.086086 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.188050 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.257172 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-nct88"] Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.257531 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerName="dnsmasq-dns" containerID="cri-o://d5d399bc9afb03fcd324b2bce9717cc0cfb5a00d78e5c367448eb09a4d1da809" gracePeriod=10 Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.380275 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.180:8776/healthcheck\": dial tcp 10.217.0.180:8776: connect: connection refused" Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.393862 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.909551 4894 generic.go:334] "Generic (PLEG): container finished" podID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerID="d5d399bc9afb03fcd324b2bce9717cc0cfb5a00d78e5c367448eb09a4d1da809" exitCode=0 Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.909766 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" event={"ID":"7de86a55-00ea-44a6-aa0e-a9687e4fd56e","Type":"ContainerDied","Data":"d5d399bc9afb03fcd324b2bce9717cc0cfb5a00d78e5c367448eb09a4d1da809"} Dec 10 19:17:30 crc kubenswrapper[4894]: I1210 19:17:30.968999 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:31 crc kubenswrapper[4894]: I1210 19:17:31.276419 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:31 crc kubenswrapper[4894]: I1210 19:17:31.740629 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:31 crc kubenswrapper[4894]: I1210 19:17:31.933651 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="cinder-scheduler" containerID="cri-o://73056b57e61fc801fe826cab24b6bd922c3a3dc33afb22e4a6f4d692fedfb6b6" gracePeriod=30 Dec 10 19:17:31 crc kubenswrapper[4894]: I1210 19:17:31.933978 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="probe" containerID="cri-o://8451a2a62025ae10893821b5dd1222630c84b53b7f4687d3ca2aeddd92b3e872" gracePeriod=30 Dec 10 19:17:32 crc kubenswrapper[4894]: I1210 19:17:32.944092 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerID="8451a2a62025ae10893821b5dd1222630c84b53b7f4687d3ca2aeddd92b3e872" exitCode=0 Dec 10 19:17:32 crc kubenswrapper[4894]: I1210 19:17:32.944167 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9b5f48b8-38ba-4884-b65a-46decaca2122","Type":"ContainerDied","Data":"8451a2a62025ae10893821b5dd1222630c84b53b7f4687d3ca2aeddd92b3e872"} Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.158756 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.165889 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299061 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-swift-storage-0\") pod \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299116 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-nb\") pod \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299162 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-combined-ca-bundle\") pod \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299226 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-config-data\") pod \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299321 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-svc\") pod \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299393 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-certs\") pod \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299420 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-sb\") pod \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299453 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fdzw\" (UniqueName: \"kubernetes.io/projected/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-kube-api-access-8fdzw\") pod \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299520 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-scripts\") pod \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299554 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-config\") pod \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\" (UID: \"7de86a55-00ea-44a6-aa0e-a9687e4fd56e\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.299608 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnb2h\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-kube-api-access-lnb2h\") pod \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\" (UID: \"bc3ed2c8-9e2c-443b-bc22-a39397986e13\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.311838 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-kube-api-access-lnb2h" (OuterVolumeSpecName: "kube-api-access-lnb2h") pod "bc3ed2c8-9e2c-443b-bc22-a39397986e13" (UID: "bc3ed2c8-9e2c-443b-bc22-a39397986e13"). InnerVolumeSpecName "kube-api-access-lnb2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.317072 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-certs" (OuterVolumeSpecName: "certs") pod "bc3ed2c8-9e2c-443b-bc22-a39397986e13" (UID: "bc3ed2c8-9e2c-443b-bc22-a39397986e13"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.317181 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-scripts" (OuterVolumeSpecName: "scripts") pod "bc3ed2c8-9e2c-443b-bc22-a39397986e13" (UID: "bc3ed2c8-9e2c-443b-bc22-a39397986e13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.319962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-kube-api-access-8fdzw" (OuterVolumeSpecName: "kube-api-access-8fdzw") pod "7de86a55-00ea-44a6-aa0e-a9687e4fd56e" (UID: "7de86a55-00ea-44a6-aa0e-a9687e4fd56e"). InnerVolumeSpecName "kube-api-access-8fdzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.371245 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc3ed2c8-9e2c-443b-bc22-a39397986e13" (UID: "bc3ed2c8-9e2c-443b-bc22-a39397986e13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.401744 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.401775 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.401785 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fdzw\" (UniqueName: \"kubernetes.io/projected/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-kube-api-access-8fdzw\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.401796 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.401804 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnb2h\" (UniqueName: \"kubernetes.io/projected/bc3ed2c8-9e2c-443b-bc22-a39397986e13-kube-api-access-lnb2h\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.404197 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-config-data" (OuterVolumeSpecName: "config-data") pod "bc3ed2c8-9e2c-443b-bc22-a39397986e13" (UID: "bc3ed2c8-9e2c-443b-bc22-a39397986e13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.412531 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7de86a55-00ea-44a6-aa0e-a9687e4fd56e" (UID: "7de86a55-00ea-44a6-aa0e-a9687e4fd56e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.418780 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7de86a55-00ea-44a6-aa0e-a9687e4fd56e" (UID: "7de86a55-00ea-44a6-aa0e-a9687e4fd56e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.433349 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7de86a55-00ea-44a6-aa0e-a9687e4fd56e" (UID: "7de86a55-00ea-44a6-aa0e-a9687e4fd56e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.447326 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7de86a55-00ea-44a6-aa0e-a9687e4fd56e" (UID: "7de86a55-00ea-44a6-aa0e-a9687e4fd56e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.470621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-config" (OuterVolumeSpecName: "config") pod "7de86a55-00ea-44a6-aa0e-a9687e4fd56e" (UID: "7de86a55-00ea-44a6-aa0e-a9687e4fd56e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.503402 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3ed2c8-9e2c-443b-bc22-a39397986e13-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.503442 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.503452 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.503463 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.503472 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.503481 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de86a55-00ea-44a6-aa0e-a9687e4fd56e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.565199 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.644688 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54c578ccd6-gksvt"] Dec 10 19:17:33 crc kubenswrapper[4894]: W1210 19:17:33.648478 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod585f9d29_8845_45a9_9dd3_8fba7042a4e3.slice/crio-469c547754c1f2caafdd08abc98d1f179e7d86ace4231d7ec2e2865d55362498 WatchSource:0}: Error finding container 469c547754c1f2caafdd08abc98d1f179e7d86ace4231d7ec2e2865d55362498: Status 404 returned error can't find the container with id 469c547754c1f2caafdd08abc98d1f179e7d86ace4231d7ec2e2865d55362498 Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.706631 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c77e8f-d949-45af-9cbb-725cfe347782-logs\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.706711 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1c77e8f-d949-45af-9cbb-725cfe347782-etc-machine-id\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.706805 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-combined-ca-bundle\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.706833 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data-custom\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.706875 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-scripts\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.707024 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.707119 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p245\" (UniqueName: \"kubernetes.io/projected/c1c77e8f-d949-45af-9cbb-725cfe347782-kube-api-access-2p245\") pod \"c1c77e8f-d949-45af-9cbb-725cfe347782\" (UID: \"c1c77e8f-d949-45af-9cbb-725cfe347782\") " Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.709262 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1c77e8f-d949-45af-9cbb-725cfe347782-logs" (OuterVolumeSpecName: "logs") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.709321 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1c77e8f-d949-45af-9cbb-725cfe347782-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.716033 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-scripts" (OuterVolumeSpecName: "scripts") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.716122 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1c77e8f-d949-45af-9cbb-725cfe347782-kube-api-access-2p245" (OuterVolumeSpecName: "kube-api-access-2p245") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "kube-api-access-2p245". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.716260 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.757620 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.789400 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data" (OuterVolumeSpecName: "config-data") pod "c1c77e8f-d949-45af-9cbb-725cfe347782" (UID: "c1c77e8f-d949-45af-9cbb-725cfe347782"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809366 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809408 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p245\" (UniqueName: \"kubernetes.io/projected/c1c77e8f-d949-45af-9cbb-725cfe347782-kube-api-access-2p245\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809427 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1c77e8f-d949-45af-9cbb-725cfe347782-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809445 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c1c77e8f-d949-45af-9cbb-725cfe347782-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809462 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809476 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.809491 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1c77e8f-d949-45af-9cbb-725cfe347782-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.962398 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54c578ccd6-gksvt" event={"ID":"585f9d29-8845-45a9-9dd3-8fba7042a4e3","Type":"ContainerStarted","Data":"abad576e683b203ed78e1076caeba163ae31f64aada0df5794443cca9f9166aa"} Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.962441 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54c578ccd6-gksvt" event={"ID":"585f9d29-8845-45a9-9dd3-8fba7042a4e3","Type":"ContainerStarted","Data":"469c547754c1f2caafdd08abc98d1f179e7d86ace4231d7ec2e2865d55362498"} Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.964113 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-vprjr" event={"ID":"bc3ed2c8-9e2c-443b-bc22-a39397986e13","Type":"ContainerDied","Data":"a2cae11379f21211665785f92b7b4967e4857e5569d8c57526144881ac66c22d"} Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.964170 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2cae11379f21211665785f92b7b4967e4857e5569d8c57526144881ac66c22d" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.964133 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-vprjr" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.978678 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" event={"ID":"7de86a55-00ea-44a6-aa0e-a9687e4fd56e","Type":"ContainerDied","Data":"9a34d78b6f4660824d19f02e3ae0ca712d353b9658d28cab6615690b6d24eede"} Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.978740 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-nct88" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.978756 4894 scope.go:117] "RemoveContainer" containerID="d5d399bc9afb03fcd324b2bce9717cc0cfb5a00d78e5c367448eb09a4d1da809" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.983793 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerStarted","Data":"f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0"} Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.984239 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-central-agent" containerID="cri-o://376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42" gracePeriod=30 Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.984423 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="sg-core" containerID="cri-o://f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95" gracePeriod=30 Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.984435 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="proxy-httpd" containerID="cri-o://f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0" gracePeriod=30 Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.984382 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:17:33 crc kubenswrapper[4894]: I1210 19:17:33.984574 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-notification-agent" containerID="cri-o://caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50" gracePeriod=30 Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:33.993408 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c1c77e8f-d949-45af-9cbb-725cfe347782","Type":"ContainerDied","Data":"c6aab3d4838cbb48d8b99f8d4615b1b811d8e19db55c07d05037cbb3d6ad8a1b"} Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:33.993524 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.005816 4894 scope.go:117] "RemoveContainer" containerID="67185d51fb7171ecf915a0e5cc1444cb49c59f5f6bf106aa30de576a4f77fd7f" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.032991 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.036205354 podStartE2EDuration="1m11.032966853s" podCreationTimestamp="2025-12-10 19:16:23 +0000 UTC" firstStartedPulling="2025-12-10 19:16:25.068313745 +0000 UTC m=+1267.863161537" lastFinishedPulling="2025-12-10 19:17:33.065075244 +0000 UTC m=+1335.859923036" observedRunningTime="2025-12-10 19:17:34.009128248 +0000 UTC m=+1336.803976050" watchObservedRunningTime="2025-12-10 19:17:34.032966853 +0000 UTC m=+1336.827814655" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.054627 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-nct88"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.065959 4894 scope.go:117] "RemoveContainer" containerID="7afe097e6035d2a0dcbe9d6afb61d942ac43ffb9f7921b9e38583340c9b68afa" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.068363 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-nct88"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.081899 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.095976 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.132185 4894 scope.go:117] "RemoveContainer" containerID="1a554d9c70f1b174e307856c4dda22fd9f0a701829c79a5824315b02ed5d337a" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.132410 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:34 crc kubenswrapper[4894]: E1210 19:17:34.133006 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" containerName="cloudkitty-db-sync" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.133088 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" containerName="cloudkitty-db-sync" Dec 10 19:17:34 crc kubenswrapper[4894]: E1210 19:17:34.133150 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerName="dnsmasq-dns" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.133198 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerName="dnsmasq-dns" Dec 10 19:17:34 crc kubenswrapper[4894]: E1210 19:17:34.133273 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerName="init" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.133324 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerName="init" Dec 10 19:17:34 crc kubenswrapper[4894]: E1210 19:17:34.133383 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.133431 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api" Dec 10 19:17:34 crc kubenswrapper[4894]: E1210 19:17:34.133488 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api-log" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.133536 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api-log" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.134004 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.134081 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" containerName="cinder-api-log" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.134161 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" containerName="cloudkitty-db-sync" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.134223 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" containerName="dnsmasq-dns" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.135412 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.137787 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.140745 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.140939 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.144138 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232257 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-scripts\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232307 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232335 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232537 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-logs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232566 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-config-data\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.232624 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrlhd\" (UniqueName: \"kubernetes.io/projected/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-kube-api-access-qrlhd\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.313426 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-sprhg"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.315154 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.317926 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.318449 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.318449 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.318926 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-gf8xv" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.319127 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336677 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336708 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-logs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-config-data\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336793 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrlhd\" (UniqueName: \"kubernetes.io/projected/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-kube-api-access-qrlhd\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336820 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-scripts\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336839 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336879 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.336925 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.347163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-config-data\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.351670 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.352187 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.352249 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-logs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.360029 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-sprhg"] Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.361439 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.361715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-scripts\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.365287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.365812 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.380555 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrlhd\" (UniqueName: \"kubernetes.io/projected/5ccf2774-6d3c-4b78-b027-867fbf8b9af0-kube-api-access-qrlhd\") pod \"cinder-api-0\" (UID: \"5ccf2774-6d3c-4b78-b027-867fbf8b9af0\") " pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.439715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-combined-ca-bundle\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.439821 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-config-data\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.439957 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v96vx\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-kube-api-access-v96vx\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.439996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-certs\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.440073 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-scripts\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.476924 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.541715 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-combined-ca-bundle\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.541789 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-config-data\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.542133 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v96vx\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-kube-api-access-v96vx\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.542195 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-certs\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.542233 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-scripts\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.545198 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-combined-ca-bundle\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.547260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-certs\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.547365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-config-data\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.547547 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-scripts\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.562210 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v96vx\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-kube-api-access-v96vx\") pod \"cloudkitty-storageinit-sprhg\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.654127 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:34 crc kubenswrapper[4894]: I1210 19:17:34.987124 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:17:34 crc kubenswrapper[4894]: W1210 19:17:34.987180 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ccf2774_6d3c_4b78_b027_867fbf8b9af0.slice/crio-41f3ba28ead3a622807c719d3e0364ce5f0ae88515ec679dac9480468d4a553a WatchSource:0}: Error finding container 41f3ba28ead3a622807c719d3e0364ce5f0ae88515ec679dac9480468d4a553a: Status 404 returned error can't find the container with id 41f3ba28ead3a622807c719d3e0364ce5f0ae88515ec679dac9480468d4a553a Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.014068 4894 generic.go:334] "Generic (PLEG): container finished" podID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerID="f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0" exitCode=0 Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.014102 4894 generic.go:334] "Generic (PLEG): container finished" podID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerID="f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95" exitCode=2 Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.014109 4894 generic.go:334] "Generic (PLEG): container finished" podID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerID="376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42" exitCode=0 Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.014146 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerDied","Data":"f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0"} Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.014172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerDied","Data":"f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95"} Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.014182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerDied","Data":"376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42"} Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.016475 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ccf2774-6d3c-4b78-b027-867fbf8b9af0","Type":"ContainerStarted","Data":"41f3ba28ead3a622807c719d3e0364ce5f0ae88515ec679dac9480468d4a553a"} Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.017664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54c578ccd6-gksvt" event={"ID":"585f9d29-8845-45a9-9dd3-8fba7042a4e3","Type":"ContainerStarted","Data":"2ed20e14df64b1e1c9db16f2e2f2bbeeab3f4ac6948462aa4c20b33d45c93adc"} Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.018815 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.018878 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.060147 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54c578ccd6-gksvt" podStartSLOduration=8.060128732 podStartE2EDuration="8.060128732s" podCreationTimestamp="2025-12-10 19:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:35.048157663 +0000 UTC m=+1337.843005455" watchObservedRunningTime="2025-12-10 19:17:35.060128732 +0000 UTC m=+1337.854976534" Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.196350 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-sprhg"] Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.554519 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7de86a55-00ea-44a6-aa0e-a9687e4fd56e" path="/var/lib/kubelet/pods/7de86a55-00ea-44a6-aa0e-a9687e4fd56e/volumes" Dec 10 19:17:35 crc kubenswrapper[4894]: I1210 19:17:35.555605 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1c77e8f-d949-45af-9cbb-725cfe347782" path="/var/lib/kubelet/pods/c1c77e8f-d949-45af-9cbb-725cfe347782/volumes" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.047318 4894 generic.go:334] "Generic (PLEG): container finished" podID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerID="dc540885ef8c2396c55fb45b6eed1e10b5a68ef93cd78c420144ae3fff163040" exitCode=0 Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.047471 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8659fd4574-zvfv8" event={"ID":"aadf98aa-0c6c-4d24-a847-e016114e615c","Type":"ContainerDied","Data":"dc540885ef8c2396c55fb45b6eed1e10b5a68ef93cd78c420144ae3fff163040"} Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.051159 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-sprhg" event={"ID":"b36157d9-6d04-40f0-a303-0ff6853ac239","Type":"ContainerStarted","Data":"2a56d533fda516fbd7e25da758f2323d2e62776ec8b4ad563cdd351c18f78582"} Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.051242 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-sprhg" event={"ID":"b36157d9-6d04-40f0-a303-0ff6853ac239","Type":"ContainerStarted","Data":"65ad1098a58d67a908dee5e6bfd9976b998eaccea26fe5813c6368db193a038a"} Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.053648 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ccf2774-6d3c-4b78-b027-867fbf8b9af0","Type":"ContainerStarted","Data":"0a2b2173d730c4af2f91e0119271487668cc5f92229383525dd25942bf4a05f5"} Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.056067 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerID="73056b57e61fc801fe826cab24b6bd922c3a3dc33afb22e4a6f4d692fedfb6b6" exitCode=0 Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.056586 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9b5f48b8-38ba-4884-b65a-46decaca2122","Type":"ContainerDied","Data":"73056b57e61fc801fe826cab24b6bd922c3a3dc33afb22e4a6f4d692fedfb6b6"} Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.082265 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-sprhg" podStartSLOduration=2.082247997 podStartE2EDuration="2.082247997s" podCreationTimestamp="2025-12-10 19:17:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:36.070779191 +0000 UTC m=+1338.865626993" watchObservedRunningTime="2025-12-10 19:17:36.082247997 +0000 UTC m=+1338.877095779" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.118179 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.294887 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data\") pod \"9b5f48b8-38ba-4884-b65a-46decaca2122\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.294969 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data-custom\") pod \"9b5f48b8-38ba-4884-b65a-46decaca2122\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.295013 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-combined-ca-bundle\") pod \"9b5f48b8-38ba-4884-b65a-46decaca2122\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.295094 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-scripts\") pod \"9b5f48b8-38ba-4884-b65a-46decaca2122\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.295169 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4tp5\" (UniqueName: \"kubernetes.io/projected/9b5f48b8-38ba-4884-b65a-46decaca2122-kube-api-access-t4tp5\") pod \"9b5f48b8-38ba-4884-b65a-46decaca2122\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.295247 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b5f48b8-38ba-4884-b65a-46decaca2122-etc-machine-id\") pod \"9b5f48b8-38ba-4884-b65a-46decaca2122\" (UID: \"9b5f48b8-38ba-4884-b65a-46decaca2122\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.295819 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b5f48b8-38ba-4884-b65a-46decaca2122-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9b5f48b8-38ba-4884-b65a-46decaca2122" (UID: "9b5f48b8-38ba-4884-b65a-46decaca2122"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.345235 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9b5f48b8-38ba-4884-b65a-46decaca2122" (UID: "9b5f48b8-38ba-4884-b65a-46decaca2122"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.345337 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-scripts" (OuterVolumeSpecName: "scripts") pod "9b5f48b8-38ba-4884-b65a-46decaca2122" (UID: "9b5f48b8-38ba-4884-b65a-46decaca2122"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.345424 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5f48b8-38ba-4884-b65a-46decaca2122-kube-api-access-t4tp5" (OuterVolumeSpecName: "kube-api-access-t4tp5") pod "9b5f48b8-38ba-4884-b65a-46decaca2122" (UID: "9b5f48b8-38ba-4884-b65a-46decaca2122"). InnerVolumeSpecName "kube-api-access-t4tp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.398014 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.398357 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.398368 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4tp5\" (UniqueName: \"kubernetes.io/projected/9b5f48b8-38ba-4884-b65a-46decaca2122-kube-api-access-t4tp5\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.398379 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b5f48b8-38ba-4884-b65a-46decaca2122-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.399902 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b5f48b8-38ba-4884-b65a-46decaca2122" (UID: "9b5f48b8-38ba-4884-b65a-46decaca2122"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.465186 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data" (OuterVolumeSpecName: "config-data") pod "9b5f48b8-38ba-4884-b65a-46decaca2122" (UID: "9b5f48b8-38ba-4884-b65a-46decaca2122"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.499685 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.499713 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5f48b8-38ba-4884-b65a-46decaca2122-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.533305 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.703222 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-ovndb-tls-certs\") pod \"aadf98aa-0c6c-4d24-a847-e016114e615c\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.703569 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7lb7\" (UniqueName: \"kubernetes.io/projected/aadf98aa-0c6c-4d24-a847-e016114e615c-kube-api-access-q7lb7\") pod \"aadf98aa-0c6c-4d24-a847-e016114e615c\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.703632 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-config\") pod \"aadf98aa-0c6c-4d24-a847-e016114e615c\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.703660 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-combined-ca-bundle\") pod \"aadf98aa-0c6c-4d24-a847-e016114e615c\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.703799 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-httpd-config\") pod \"aadf98aa-0c6c-4d24-a847-e016114e615c\" (UID: \"aadf98aa-0c6c-4d24-a847-e016114e615c\") " Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.711531 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aadf98aa-0c6c-4d24-a847-e016114e615c-kube-api-access-q7lb7" (OuterVolumeSpecName: "kube-api-access-q7lb7") pod "aadf98aa-0c6c-4d24-a847-e016114e615c" (UID: "aadf98aa-0c6c-4d24-a847-e016114e615c"). InnerVolumeSpecName "kube-api-access-q7lb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.714960 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "aadf98aa-0c6c-4d24-a847-e016114e615c" (UID: "aadf98aa-0c6c-4d24-a847-e016114e615c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.762912 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aadf98aa-0c6c-4d24-a847-e016114e615c" (UID: "aadf98aa-0c6c-4d24-a847-e016114e615c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.785165 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-config" (OuterVolumeSpecName: "config") pod "aadf98aa-0c6c-4d24-a847-e016114e615c" (UID: "aadf98aa-0c6c-4d24-a847-e016114e615c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.790737 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "aadf98aa-0c6c-4d24-a847-e016114e615c" (UID: "aadf98aa-0c6c-4d24-a847-e016114e615c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.807072 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.807100 4894 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.807113 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7lb7\" (UniqueName: \"kubernetes.io/projected/aadf98aa-0c6c-4d24-a847-e016114e615c-kube-api-access-q7lb7\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.807121 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:36 crc kubenswrapper[4894]: I1210 19:17:36.807129 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aadf98aa-0c6c-4d24-a847-e016114e615c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.067122 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ccf2774-6d3c-4b78-b027-867fbf8b9af0","Type":"ContainerStarted","Data":"e184c94a600fd9a7b15fb2d2b31648fe7eb566c3213ef999fd21cdc9b5b632be"} Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.068239 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.069375 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9b5f48b8-38ba-4884-b65a-46decaca2122","Type":"ContainerDied","Data":"6c1a9ac7b1fd4859c3fe52ac131b451f33e29a8c128dafd87bdab517b955a720"} Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.069405 4894 scope.go:117] "RemoveContainer" containerID="8451a2a62025ae10893821b5dd1222630c84b53b7f4687d3ca2aeddd92b3e872" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.069533 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.074488 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8659fd4574-zvfv8" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.074880 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8659fd4574-zvfv8" event={"ID":"aadf98aa-0c6c-4d24-a847-e016114e615c","Type":"ContainerDied","Data":"fc765c78b133ef20d11652b28d0271dce3045a031cb39c416e4fa849e641a250"} Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.105704 4894 scope.go:117] "RemoveContainer" containerID="73056b57e61fc801fe826cab24b6bd922c3a3dc33afb22e4a6f4d692fedfb6b6" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.107343 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.107325691 podStartE2EDuration="3.107325691s" podCreationTimestamp="2025-12-10 19:17:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:37.097000886 +0000 UTC m=+1339.891848768" watchObservedRunningTime="2025-12-10 19:17:37.107325691 +0000 UTC m=+1339.902173483" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.139036 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8659fd4574-zvfv8"] Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.144325 4894 scope.go:117] "RemoveContainer" containerID="0ef8686844d0e28e9c3e28d449ff88c64f666e286d4903453a871c2aeb8353c2" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.150398 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8659fd4574-zvfv8"] Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.163348 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.171012 4894 scope.go:117] "RemoveContainer" containerID="dc540885ef8c2396c55fb45b6eed1e10b5a68ef93cd78c420144ae3fff163040" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.174475 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.199941 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:37 crc kubenswrapper[4894]: E1210 19:17:37.200484 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-httpd" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200502 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-httpd" Dec 10 19:17:37 crc kubenswrapper[4894]: E1210 19:17:37.200536 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-api" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200545 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-api" Dec 10 19:17:37 crc kubenswrapper[4894]: E1210 19:17:37.200562 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="probe" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200569 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="probe" Dec 10 19:17:37 crc kubenswrapper[4894]: E1210 19:17:37.200587 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="cinder-scheduler" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200594 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="cinder-scheduler" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200818 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-httpd" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200836 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="cinder-scheduler" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200863 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" containerName="probe" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.200881 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" containerName="neutron-api" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.202358 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.204841 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.217984 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.318367 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-config-data\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.318730 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-scripts\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.318906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.318989 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.319356 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de91d530-228b-482b-baf4-b7b4b1dd6fec-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.319480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv97s\" (UniqueName: \"kubernetes.io/projected/de91d530-228b-482b-baf4-b7b4b1dd6fec-kube-api-access-rv97s\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422003 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de91d530-228b-482b-baf4-b7b4b1dd6fec-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv97s\" (UniqueName: \"kubernetes.io/projected/de91d530-228b-482b-baf4-b7b4b1dd6fec-kube-api-access-rv97s\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422116 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-config-data\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422141 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-scripts\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422150 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de91d530-228b-482b-baf4-b7b4b1dd6fec-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422362 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.422404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.426555 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.426970 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-scripts\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.427576 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.427922 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de91d530-228b-482b-baf4-b7b4b1dd6fec-config-data\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.442449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv97s\" (UniqueName: \"kubernetes.io/projected/de91d530-228b-482b-baf4-b7b4b1dd6fec-kube-api-access-rv97s\") pod \"cinder-scheduler-0\" (UID: \"de91d530-228b-482b-baf4-b7b4b1dd6fec\") " pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.561414 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.561711 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b5f48b8-38ba-4884-b65a-46decaca2122" path="/var/lib/kubelet/pods/9b5f48b8-38ba-4884-b65a-46decaca2122/volumes" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.562334 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aadf98aa-0c6c-4d24-a847-e016114e615c" path="/var/lib/kubelet/pods/aadf98aa-0c6c-4d24-a847-e016114e615c/volumes" Dec 10 19:17:37 crc kubenswrapper[4894]: I1210 19:17:37.886413 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035376 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-scripts\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-config-data\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035585 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-combined-ca-bundle\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035620 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-sg-core-conf-yaml\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035678 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dggnm\" (UniqueName: \"kubernetes.io/projected/fe2420b5-49dd-45eb-875e-81014391d3a0-kube-api-access-dggnm\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035775 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-run-httpd\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.035822 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-log-httpd\") pod \"fe2420b5-49dd-45eb-875e-81014391d3a0\" (UID: \"fe2420b5-49dd-45eb-875e-81014391d3a0\") " Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.037801 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.037868 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.042407 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-scripts" (OuterVolumeSpecName: "scripts") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.042468 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe2420b5-49dd-45eb-875e-81014391d3a0-kube-api-access-dggnm" (OuterVolumeSpecName: "kube-api-access-dggnm") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "kube-api-access-dggnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.073055 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.084641 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.131519 4894 generic.go:334] "Generic (PLEG): container finished" podID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerID="caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50" exitCode=0 Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.132021 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.133525 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerDied","Data":"caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50"} Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.133634 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe2420b5-49dd-45eb-875e-81014391d3a0","Type":"ContainerDied","Data":"d1d39bc215c37245dedacce0112e5e2c43d0dbcf6f4b7ed36077a475467c8cb5"} Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.133674 4894 scope.go:117] "RemoveContainer" containerID="f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.147091 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.147135 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe2420b5-49dd-45eb-875e-81014391d3a0-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.147148 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.147162 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.147178 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dggnm\" (UniqueName: \"kubernetes.io/projected/fe2420b5-49dd-45eb-875e-81014391d3a0-kube-api-access-dggnm\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.150987 4894 generic.go:334] "Generic (PLEG): container finished" podID="b36157d9-6d04-40f0-a303-0ff6853ac239" containerID="2a56d533fda516fbd7e25da758f2323d2e62776ec8b4ad563cdd351c18f78582" exitCode=0 Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.151927 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-sprhg" event={"ID":"b36157d9-6d04-40f0-a303-0ff6853ac239","Type":"ContainerDied","Data":"2a56d533fda516fbd7e25da758f2323d2e62776ec8b4ad563cdd351c18f78582"} Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.162086 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.182113 4894 scope.go:117] "RemoveContainer" containerID="f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.188793 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-config-data" (OuterVolumeSpecName: "config-data") pod "fe2420b5-49dd-45eb-875e-81014391d3a0" (UID: "fe2420b5-49dd-45eb-875e-81014391d3a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.249726 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.250291 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2420b5-49dd-45eb-875e-81014391d3a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.323406 4894 scope.go:117] "RemoveContainer" containerID="caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.377232 4894 scope.go:117] "RemoveContainer" containerID="376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.514015 4894 scope.go:117] "RemoveContainer" containerID="f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.520124 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.525012 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0\": container with ID starting with f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0 not found: ID does not exist" containerID="f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.525057 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0"} err="failed to get container status \"f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0\": rpc error: code = NotFound desc = could not find container \"f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0\": container with ID starting with f6b6b952287d655eb6254780e2284a179fcbd2f7cc21c928b15d37976145ebd0 not found: ID does not exist" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.525083 4894 scope.go:117] "RemoveContainer" containerID="f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.534976 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95\": container with ID starting with f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95 not found: ID does not exist" containerID="f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.535023 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95"} err="failed to get container status \"f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95\": rpc error: code = NotFound desc = could not find container \"f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95\": container with ID starting with f80db52dc98c284936f56cc53dbd4cb7b07175c45473d5ad598e0d80a88efc95 not found: ID does not exist" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.535050 4894 scope.go:117] "RemoveContainer" containerID="caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.564013 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50\": container with ID starting with caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50 not found: ID does not exist" containerID="caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.564071 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50"} err="failed to get container status \"caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50\": rpc error: code = NotFound desc = could not find container \"caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50\": container with ID starting with caee45ca734b83071224465da900b7e071aee3c99d6728d624ae6ab071b46f50 not found: ID does not exist" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.564104 4894 scope.go:117] "RemoveContainer" containerID="376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.567972 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42\": container with ID starting with 376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42 not found: ID does not exist" containerID="376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.568006 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42"} err="failed to get container status \"376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42\": rpc error: code = NotFound desc = could not find container \"376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42\": container with ID starting with 376db3cad3bbdf2b276dfe4edda94436343a53fee8b35a860a548cea61e27f42 not found: ID does not exist" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.572906 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.592768 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.593303 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="proxy-httpd" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.593318 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="proxy-httpd" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.593329 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="sg-core" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.593336 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="sg-core" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.593363 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-central-agent" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.593372 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-central-agent" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.593395 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-notification-agent" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.593402 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-notification-agent" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.607022 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-central-agent" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.607079 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="proxy-httpd" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.607101 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="ceilometer-notification-agent" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.607118 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" containerName="sg-core" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.609524 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.609644 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.618173 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.618614 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677102 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-run-httpd\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677174 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677199 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-scripts\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677233 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677308 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-log-httpd\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677331 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-config-data\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.677381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bcwl\" (UniqueName: \"kubernetes.io/projected/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-kube-api-access-8bcwl\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.718091 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.778664 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-run-httpd\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.778751 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.778783 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-scripts\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.778817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.778923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-log-httpd\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.778957 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-config-data\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.779005 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bcwl\" (UniqueName: \"kubernetes.io/projected/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-kube-api-access-8bcwl\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.779419 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-run-httpd\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.779693 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-log-httpd\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.783493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.791584 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.793573 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-scripts\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.794252 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-config-data\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.803723 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bcwl\" (UniqueName: \"kubernetes.io/projected/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-kube-api-access-8bcwl\") pod \"ceilometer-0\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " pod="openstack/ceilometer-0" Dec 10 19:17:38 crc kubenswrapper[4894]: E1210 19:17:38.819462 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe2420b5_49dd_45eb_875e_81014391d3a0.slice/crio-d1d39bc215c37245dedacce0112e5e2c43d0dbcf6f4b7ed36077a475467c8cb5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe2420b5_49dd_45eb_875e_81014391d3a0.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.888756 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7ff967f45b-c9flb" Dec 10 19:17:38 crc kubenswrapper[4894]: I1210 19:17:38.951710 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.174136 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"de91d530-228b-482b-baf4-b7b4b1dd6fec","Type":"ContainerStarted","Data":"fd454e85d78e887abd5ade825e14344872151bb9b786a15fcffce34c1ad36cc2"} Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.174538 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"de91d530-228b-482b-baf4-b7b4b1dd6fec","Type":"ContainerStarted","Data":"b012540bce740c95f4ebbff10fc7c2614bf44fbde26e0f46c34e23d9b41f568d"} Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.573617 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe2420b5-49dd-45eb-875e-81014391d3a0" path="/var/lib/kubelet/pods/fe2420b5-49dd-45eb-875e-81014391d3a0/volumes" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.574770 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.815170 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.917028 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-certs\") pod \"b36157d9-6d04-40f0-a303-0ff6853ac239\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.917193 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-combined-ca-bundle\") pod \"b36157d9-6d04-40f0-a303-0ff6853ac239\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.917302 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-scripts\") pod \"b36157d9-6d04-40f0-a303-0ff6853ac239\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.917334 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v96vx\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-kube-api-access-v96vx\") pod \"b36157d9-6d04-40f0-a303-0ff6853ac239\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.917388 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-config-data\") pod \"b36157d9-6d04-40f0-a303-0ff6853ac239\" (UID: \"b36157d9-6d04-40f0-a303-0ff6853ac239\") " Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.927003 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-scripts" (OuterVolumeSpecName: "scripts") pod "b36157d9-6d04-40f0-a303-0ff6853ac239" (UID: "b36157d9-6d04-40f0-a303-0ff6853ac239"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.927065 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-certs" (OuterVolumeSpecName: "certs") pod "b36157d9-6d04-40f0-a303-0ff6853ac239" (UID: "b36157d9-6d04-40f0-a303-0ff6853ac239"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.927096 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-kube-api-access-v96vx" (OuterVolumeSpecName: "kube-api-access-v96vx") pod "b36157d9-6d04-40f0-a303-0ff6853ac239" (UID: "b36157d9-6d04-40f0-a303-0ff6853ac239"). InnerVolumeSpecName "kube-api-access-v96vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.951038 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-config-data" (OuterVolumeSpecName: "config-data") pod "b36157d9-6d04-40f0-a303-0ff6853ac239" (UID: "b36157d9-6d04-40f0-a303-0ff6853ac239"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:39 crc kubenswrapper[4894]: I1210 19:17:39.957990 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b36157d9-6d04-40f0-a303-0ff6853ac239" (UID: "b36157d9-6d04-40f0-a303-0ff6853ac239"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.020669 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.020704 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v96vx\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-kube-api-access-v96vx\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.020715 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.020724 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b36157d9-6d04-40f0-a303-0ff6853ac239-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.020732 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36157d9-6d04-40f0-a303-0ff6853ac239-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.188275 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-sprhg" event={"ID":"b36157d9-6d04-40f0-a303-0ff6853ac239","Type":"ContainerDied","Data":"65ad1098a58d67a908dee5e6bfd9976b998eaccea26fe5813c6368db193a038a"} Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.188313 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65ad1098a58d67a908dee5e6bfd9976b998eaccea26fe5813c6368db193a038a" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.188798 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-sprhg" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.190577 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"de91d530-228b-482b-baf4-b7b4b1dd6fec","Type":"ContainerStarted","Data":"d4e3d461a81b416f98a467e1859b3d920af2f2b912d348322569bcc09a8b634b"} Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.191735 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerStarted","Data":"b9b0b9a6a3c5d0ae70b9c9239f643cf5b5b757048cca3dbb097674f6bd518ed4"} Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.248364 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.248342428 podStartE2EDuration="3.248342428s" podCreationTimestamp="2025-12-10 19:17:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:40.217390403 +0000 UTC m=+1343.012238205" watchObservedRunningTime="2025-12-10 19:17:40.248342428 +0000 UTC m=+1343.043190220" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.367393 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:40 crc kubenswrapper[4894]: E1210 19:17:40.368625 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36157d9-6d04-40f0-a303-0ff6853ac239" containerName="cloudkitty-storageinit" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.368641 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36157d9-6d04-40f0-a303-0ff6853ac239" containerName="cloudkitty-storageinit" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.369692 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36157d9-6d04-40f0-a303-0ff6853ac239" containerName="cloudkitty-storageinit" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.372724 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.374374 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.374926 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-gf8xv" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.375182 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.375255 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.378916 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.387791 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.434915 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg7tz\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-kube-api-access-qg7tz\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.434978 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-scripts\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.435008 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.435048 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.435090 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-certs\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.435129 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.465871 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86d9875b97-pnsfc"] Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.467670 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.477430 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86d9875b97-pnsfc"] Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.537306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-certs\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538139 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-config\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538280 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-nb\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538407 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-swift-storage-0\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538491 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m7zv\" (UniqueName: \"kubernetes.io/projected/b4ef5729-6000-469a-bf66-a53c71bda7e0-kube-api-access-2m7zv\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538572 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg7tz\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-kube-api-access-qg7tz\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538663 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-svc\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538701 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-scripts\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-sb\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538777 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.538923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.544393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.547236 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-scripts\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.551350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.554918 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.569360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-certs\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.577564 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg7tz\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-kube-api-access-qg7tz\") pod \"cloudkitty-proc-0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.610401 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.615103 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.617754 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.630732 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.641946 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m7zv\" (UniqueName: \"kubernetes.io/projected/b4ef5729-6000-469a-bf66-a53c71bda7e0-kube-api-access-2m7zv\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.642069 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-svc\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.642121 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-sb\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.642911 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-config\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.643005 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-nb\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.643110 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-swift-storage-0\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.645442 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-svc\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.645623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-sb\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.646186 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-config\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.646390 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-nb\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.665569 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-swift-storage-0\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.686453 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m7zv\" (UniqueName: \"kubernetes.io/projected/b4ef5729-6000-469a-bf66-a53c71bda7e0-kube-api-access-2m7zv\") pod \"dnsmasq-dns-86d9875b97-pnsfc\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.705668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758082 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758228 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72ba0006-fdbc-4330-9d5f-bbfa5decf545-logs\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758289 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8cxc\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-kube-api-access-q8cxc\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-certs\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758478 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-scripts\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.758570 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.798170 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.859967 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-certs\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.860350 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-scripts\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.860383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.860407 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.860453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72ba0006-fdbc-4330-9d5f-bbfa5decf545-logs\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.860486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.860505 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8cxc\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-kube-api-access-q8cxc\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.865222 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72ba0006-fdbc-4330-9d5f-bbfa5decf545-logs\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.865365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-certs\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.875059 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.878930 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.879290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-scripts\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.884326 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:40 crc kubenswrapper[4894]: I1210 19:17:40.891441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8cxc\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-kube-api-access-q8cxc\") pod \"cloudkitty-api-0\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:41 crc kubenswrapper[4894]: I1210 19:17:41.142497 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:17:41 crc kubenswrapper[4894]: I1210 19:17:41.563385 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:41 crc kubenswrapper[4894]: W1210 19:17:41.568984 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67b548c7_9e9f_4c17_89c7_96df0a6670b0.slice/crio-208c1699dee0180c01ec259acac9adbf10813b1d5afe010aab9e8347ed59237f WatchSource:0}: Error finding container 208c1699dee0180c01ec259acac9adbf10813b1d5afe010aab9e8347ed59237f: Status 404 returned error can't find the container with id 208c1699dee0180c01ec259acac9adbf10813b1d5afe010aab9e8347ed59237f Dec 10 19:17:41 crc kubenswrapper[4894]: I1210 19:17:41.714875 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86d9875b97-pnsfc"] Dec 10 19:17:41 crc kubenswrapper[4894]: I1210 19:17:41.859349 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.236088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"72ba0006-fdbc-4330-9d5f-bbfa5decf545","Type":"ContainerStarted","Data":"bf26864ee39cae57e5fcab67a75c02b0a0f9ecff7ffae67b6557c2610c66368d"} Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.236336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"72ba0006-fdbc-4330-9d5f-bbfa5decf545","Type":"ContainerStarted","Data":"a28c6ae30905189b5934f4a37795a1667da13508dc872421e3a0fcbfc008c455"} Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.238040 4894 generic.go:334] "Generic (PLEG): container finished" podID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerID="13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0" exitCode=0 Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.238145 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" event={"ID":"b4ef5729-6000-469a-bf66-a53c71bda7e0","Type":"ContainerDied","Data":"13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0"} Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.238226 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" event={"ID":"b4ef5729-6000-469a-bf66-a53c71bda7e0","Type":"ContainerStarted","Data":"4fd6031d77791ddf20799fa26f085733601a0e488268d7cdb3fd19af8a98ec3e"} Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.244827 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"67b548c7-9e9f-4c17-89c7-96df0a6670b0","Type":"ContainerStarted","Data":"208c1699dee0180c01ec259acac9adbf10813b1d5afe010aab9e8347ed59237f"} Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.250760 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerStarted","Data":"2bb580e2b2ff5d45214eac5c31af25d5aa6cae324e85cfba940e4af4840047eb"} Dec 10 19:17:42 crc kubenswrapper[4894]: I1210 19:17:42.562547 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 19:17:43 crc kubenswrapper[4894]: I1210 19:17:43.266893 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"72ba0006-fdbc-4330-9d5f-bbfa5decf545","Type":"ContainerStarted","Data":"7ae3372213b38e954187c48bc245885342733013650237dea1649cadbf8237a4"} Dec 10 19:17:43 crc kubenswrapper[4894]: I1210 19:17:43.267435 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 10 19:17:43 crc kubenswrapper[4894]: I1210 19:17:43.302991 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=3.302969269 podStartE2EDuration="3.302969269s" podCreationTimestamp="2025-12-10 19:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:43.292887441 +0000 UTC m=+1346.087735233" watchObservedRunningTime="2025-12-10 19:17:43.302969269 +0000 UTC m=+1346.097817061" Dec 10 19:17:43 crc kubenswrapper[4894]: I1210 19:17:43.761996 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.033396 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.142619 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54c578ccd6-gksvt" Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.213022 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5cb6f668cd-tg9dt"] Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.213558 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5cb6f668cd-tg9dt" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api-log" containerID="cri-o://95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe" gracePeriod=30 Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.213677 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5cb6f668cd-tg9dt" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api" containerID="cri-o://ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5" gracePeriod=30 Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.290287 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" event={"ID":"b4ef5729-6000-469a-bf66-a53c71bda7e0","Type":"ContainerStarted","Data":"fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd"} Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.291514 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.293678 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"67b548c7-9e9f-4c17-89c7-96df0a6670b0","Type":"ContainerStarted","Data":"bfc993891b2abfaea1dad50a4536709dd87497086cd7b8728c9859894d362fec"} Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.298280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerStarted","Data":"588f33e6583a95c9043727441bb20f5d0f26e5a9d72e7b6be25faa790971b429"} Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.298317 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerStarted","Data":"0a41a2abefaf26050d477881a439fbe24383960e585a45e145ac3fe34a8748c7"} Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.298437 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api-log" containerID="cri-o://bf26864ee39cae57e5fcab67a75c02b0a0f9ecff7ffae67b6557c2610c66368d" gracePeriod=30 Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.298691 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api" containerID="cri-o://7ae3372213b38e954187c48bc245885342733013650237dea1649cadbf8237a4" gracePeriod=30 Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.316617 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" podStartSLOduration=5.316593983 podStartE2EDuration="5.316593983s" podCreationTimestamp="2025-12-10 19:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:45.309444432 +0000 UTC m=+1348.104292234" watchObservedRunningTime="2025-12-10 19:17:45.316593983 +0000 UTC m=+1348.111441775" Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.386031 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.346731147 podStartE2EDuration="5.386012885s" podCreationTimestamp="2025-12-10 19:17:40 +0000 UTC" firstStartedPulling="2025-12-10 19:17:41.57063135 +0000 UTC m=+1344.365479142" lastFinishedPulling="2025-12-10 19:17:43.609913088 +0000 UTC m=+1346.404760880" observedRunningTime="2025-12-10 19:17:45.335142288 +0000 UTC m=+1348.129990090" watchObservedRunningTime="2025-12-10 19:17:45.386012885 +0000 UTC m=+1348.180860687" Dec 10 19:17:45 crc kubenswrapper[4894]: I1210 19:17:45.443031 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.339074 4894 generic.go:334] "Generic (PLEG): container finished" podID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerID="95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe" exitCode=143 Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.339418 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb6f668cd-tg9dt" event={"ID":"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4","Type":"ContainerDied","Data":"95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe"} Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.401012 4894 generic.go:334] "Generic (PLEG): container finished" podID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerID="7ae3372213b38e954187c48bc245885342733013650237dea1649cadbf8237a4" exitCode=0 Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.401052 4894 generic.go:334] "Generic (PLEG): container finished" podID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerID="bf26864ee39cae57e5fcab67a75c02b0a0f9ecff7ffae67b6557c2610c66368d" exitCode=143 Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.401693 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"72ba0006-fdbc-4330-9d5f-bbfa5decf545","Type":"ContainerDied","Data":"7ae3372213b38e954187c48bc245885342733013650237dea1649cadbf8237a4"} Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.401718 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"72ba0006-fdbc-4330-9d5f-bbfa5decf545","Type":"ContainerDied","Data":"bf26864ee39cae57e5fcab67a75c02b0a0f9ecff7ffae67b6557c2610c66368d"} Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.401727 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"72ba0006-fdbc-4330-9d5f-bbfa5decf545","Type":"ContainerDied","Data":"a28c6ae30905189b5934f4a37795a1667da13508dc872421e3a0fcbfc008c455"} Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.401736 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a28c6ae30905189b5934f4a37795a1667da13508dc872421e3a0fcbfc008c455" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.448139 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.540999 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-scripts\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.541259 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72ba0006-fdbc-4330-9d5f-bbfa5decf545-logs\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.541627 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.542019 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-certs\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.542188 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-combined-ca-bundle\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.542287 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8cxc\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-kube-api-access-q8cxc\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.542403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data-custom\") pod \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\" (UID: \"72ba0006-fdbc-4330-9d5f-bbfa5decf545\") " Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.541540 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ba0006-fdbc-4330-9d5f-bbfa5decf545-logs" (OuterVolumeSpecName: "logs") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.567067 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.567111 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-certs" (OuterVolumeSpecName: "certs") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.570008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-kube-api-access-q8cxc" (OuterVolumeSpecName: "kube-api-access-q8cxc") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "kube-api-access-q8cxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.590491 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-scripts" (OuterVolumeSpecName: "scripts") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.619987 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.629105 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data" (OuterVolumeSpecName: "config-data") pod "72ba0006-fdbc-4330-9d5f-bbfa5decf545" (UID: "72ba0006-fdbc-4330-9d5f-bbfa5decf545"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648298 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648335 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648347 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648357 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8cxc\" (UniqueName: \"kubernetes.io/projected/72ba0006-fdbc-4330-9d5f-bbfa5decf545-kube-api-access-q8cxc\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648367 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648375 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ba0006-fdbc-4330-9d5f-bbfa5decf545-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:46 crc kubenswrapper[4894]: I1210 19:17:46.648382 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72ba0006-fdbc-4330-9d5f-bbfa5decf545-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.409227 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.409363 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="67b548c7-9e9f-4c17-89c7-96df0a6670b0" containerName="cloudkitty-proc" containerID="cri-o://bfc993891b2abfaea1dad50a4536709dd87497086cd7b8728c9859894d362fec" gracePeriod=30 Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.445051 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.460573 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.479430 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:47 crc kubenswrapper[4894]: E1210 19:17:47.479805 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api-log" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.479821 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api-log" Dec 10 19:17:47 crc kubenswrapper[4894]: E1210 19:17:47.479873 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.479881 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.480050 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.480070 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" containerName="cloudkitty-api-log" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.481087 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.483589 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.483717 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.485310 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.493951 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.557921 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72ba0006-fdbc-4330-9d5f-bbfa5decf545" path="/var/lib/kubelet/pods/72ba0006-fdbc-4330-9d5f-bbfa5decf545/volumes" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566039 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdftv\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-kube-api-access-cdftv\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566121 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566166 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566227 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-scripts\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566256 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566323 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566347 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-logs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.566450 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668096 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668147 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-logs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668208 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668277 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdftv\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-kube-api-access-cdftv\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668335 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668511 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-scripts\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668544 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.668558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.669410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-logs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.674013 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-scripts\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.681377 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.681954 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.682635 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.684386 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.691360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.692054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdftv\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-kube-api-access-cdftv\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.704546 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.796187 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:17:47 crc kubenswrapper[4894]: I1210 19:17:47.890307 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 19:17:48 crc kubenswrapper[4894]: I1210 19:17:48.362993 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:17:48 crc kubenswrapper[4894]: I1210 19:17:48.420051 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"38368ffb-25fe-478c-a1ab-5ca2b4a0c376","Type":"ContainerStarted","Data":"40da522b4d68dc127a9e14be52191ea63f96a20ddb4b1e6d196cd8a1c5ff5d29"} Dec 10 19:17:48 crc kubenswrapper[4894]: I1210 19:17:48.438327 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerStarted","Data":"eecc89b8810905ee7b2dd0aac8ac03c1189da0fc52168bb64ae171030487eb37"} Dec 10 19:17:48 crc kubenswrapper[4894]: I1210 19:17:48.438813 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:17:48 crc kubenswrapper[4894]: I1210 19:17:48.488222 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.847023568 podStartE2EDuration="10.488205806s" podCreationTimestamp="2025-12-10 19:17:38 +0000 UTC" firstStartedPulling="2025-12-10 19:17:39.587168691 +0000 UTC m=+1342.382016483" lastFinishedPulling="2025-12-10 19:17:47.228350929 +0000 UTC m=+1350.023198721" observedRunningTime="2025-12-10 19:17:48.486714196 +0000 UTC m=+1351.281561988" watchObservedRunningTime="2025-12-10 19:17:48.488205806 +0000 UTC m=+1351.283053598" Dec 10 19:17:48 crc kubenswrapper[4894]: I1210 19:17:48.984301 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.034576 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6bf6fc85f8-h5xlr" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.315495 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.429743 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-logs\") pod \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.429903 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data\") pod \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.429958 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-combined-ca-bundle\") pod \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.429996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4q5k\" (UniqueName: \"kubernetes.io/projected/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-kube-api-access-g4q5k\") pod \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.430074 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data-custom\") pod \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\" (UID: \"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4\") " Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.431134 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-logs" (OuterVolumeSpecName: "logs") pod "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" (UID: "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.436621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-kube-api-access-g4q5k" (OuterVolumeSpecName: "kube-api-access-g4q5k") pod "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" (UID: "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4"). InnerVolumeSpecName "kube-api-access-g4q5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.438950 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" (UID: "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.451035 4894 generic.go:334] "Generic (PLEG): container finished" podID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerID="ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5" exitCode=0 Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.451164 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5cb6f668cd-tg9dt" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.452098 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb6f668cd-tg9dt" event={"ID":"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4","Type":"ContainerDied","Data":"ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5"} Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.452160 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5cb6f668cd-tg9dt" event={"ID":"44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4","Type":"ContainerDied","Data":"408c963617aa4c7876f76d0c81605ca21274548c9b7c9513b8512ba0b0e57d0e"} Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.452181 4894 scope.go:117] "RemoveContainer" containerID="ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.480065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"38368ffb-25fe-478c-a1ab-5ca2b4a0c376","Type":"ContainerStarted","Data":"fcd772eee6a7ebb5c88934d022ebd48ecddd199bbee9c9360156d698935d92e1"} Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.480097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"38368ffb-25fe-478c-a1ab-5ca2b4a0c376","Type":"ContainerStarted","Data":"885c37f3498f8bbf498badf4e3529e8cc0959aaa9cd8b6f56b38a0a27300bf4e"} Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.480112 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.497040 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" (UID: "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.500103 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.5000856970000003 podStartE2EDuration="2.500085697s" podCreationTimestamp="2025-12-10 19:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:49.496246715 +0000 UTC m=+1352.291094507" watchObservedRunningTime="2025-12-10 19:17:49.500085697 +0000 UTC m=+1352.294933489" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.533425 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.533445 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.533458 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4q5k\" (UniqueName: \"kubernetes.io/projected/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-kube-api-access-g4q5k\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.533468 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.537944 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data" (OuterVolumeSpecName: "config-data") pod "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" (UID: "44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.563160 4894 scope.go:117] "RemoveContainer" containerID="95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.579903 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.579953 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.586111 4894 scope.go:117] "RemoveContainer" containerID="ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5" Dec 10 19:17:49 crc kubenswrapper[4894]: E1210 19:17:49.587206 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5\": container with ID starting with ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5 not found: ID does not exist" containerID="ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.587249 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5"} err="failed to get container status \"ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5\": rpc error: code = NotFound desc = could not find container \"ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5\": container with ID starting with ec74b60e769b29540dec0c5d62c5ea737d29fa02db7e0701a29fd169570ebbc5 not found: ID does not exist" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.587273 4894 scope.go:117] "RemoveContainer" containerID="95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe" Dec 10 19:17:49 crc kubenswrapper[4894]: E1210 19:17:49.590684 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe\": container with ID starting with 95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe not found: ID does not exist" containerID="95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.590729 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe"} err="failed to get container status \"95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe\": rpc error: code = NotFound desc = could not find container \"95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe\": container with ID starting with 95732cf55911b153eba6ea742c28acad668ee911f5fa3310becd096e1bf301fe not found: ID does not exist" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.635303 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.773972 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5cb6f668cd-tg9dt"] Dec 10 19:17:49 crc kubenswrapper[4894]: I1210 19:17:49.784375 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5cb6f668cd-tg9dt"] Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.253021 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 10 19:17:50 crc kubenswrapper[4894]: E1210 19:17:50.253464 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.253481 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api" Dec 10 19:17:50 crc kubenswrapper[4894]: E1210 19:17:50.253497 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api-log" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.253504 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api-log" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.253716 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.253735 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" containerName="barbican-api-log" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.254467 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.256374 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.256753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-r2fqd" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.256923 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.269250 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.347590 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcqjr\" (UniqueName: \"kubernetes.io/projected/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-kube-api-access-pcqjr\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.347650 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-openstack-config\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.347761 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-openstack-config-secret\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.347810 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.449698 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcqjr\" (UniqueName: \"kubernetes.io/projected/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-kube-api-access-pcqjr\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.449739 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-openstack-config\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.449797 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-openstack-config-secret\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.449827 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.450653 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-openstack-config\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.455324 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-openstack-config-secret\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.460126 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.469998 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcqjr\" (UniqueName: \"kubernetes.io/projected/3d4126ef-dbcf-4614-a1fe-37c7c0be7223-kube-api-access-pcqjr\") pod \"openstackclient\" (UID: \"3d4126ef-dbcf-4614-a1fe-37c7c0be7223\") " pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.574404 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.800067 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.881078 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2frqm"] Dec 10 19:17:50 crc kubenswrapper[4894]: I1210 19:17:50.881310 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerName="dnsmasq-dns" containerID="cri-o://42af34c411f9b80c3afb60391e4a65aa54154b78cc246261bf626674ae222948" gracePeriod=10 Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.140471 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.498415 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3d4126ef-dbcf-4614-a1fe-37c7c0be7223","Type":"ContainerStarted","Data":"5c4c345b69ab035ee47e8f9710b5228b71b79a5eb6544f6eadd88e216ca3fcf6"} Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.500603 4894 generic.go:334] "Generic (PLEG): container finished" podID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerID="42af34c411f9b80c3afb60391e4a65aa54154b78cc246261bf626674ae222948" exitCode=0 Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.500649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" event={"ID":"9a08e52d-bd79-4511-ab2e-86dfb4a42449","Type":"ContainerDied","Data":"42af34c411f9b80c3afb60391e4a65aa54154b78cc246261bf626674ae222948"} Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.573014 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4" path="/var/lib/kubelet/pods/44ad3e10-b7bd-4b29-b2f5-f673aea6b2f4/volumes" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.692892 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.789913 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-swift-storage-0\") pod \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.790049 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-nb\") pod \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.790079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25x9r\" (UniqueName: \"kubernetes.io/projected/9a08e52d-bd79-4511-ab2e-86dfb4a42449-kube-api-access-25x9r\") pod \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.790135 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-sb\") pod \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.790276 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-config\") pod \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.790377 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-svc\") pod \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\" (UID: \"9a08e52d-bd79-4511-ab2e-86dfb4a42449\") " Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.808066 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a08e52d-bd79-4511-ab2e-86dfb4a42449-kube-api-access-25x9r" (OuterVolumeSpecName: "kube-api-access-25x9r") pod "9a08e52d-bd79-4511-ab2e-86dfb4a42449" (UID: "9a08e52d-bd79-4511-ab2e-86dfb4a42449"). InnerVolumeSpecName "kube-api-access-25x9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.871351 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a08e52d-bd79-4511-ab2e-86dfb4a42449" (UID: "9a08e52d-bd79-4511-ab2e-86dfb4a42449"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.871734 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-config" (OuterVolumeSpecName: "config") pod "9a08e52d-bd79-4511-ab2e-86dfb4a42449" (UID: "9a08e52d-bd79-4511-ab2e-86dfb4a42449"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.872454 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a08e52d-bd79-4511-ab2e-86dfb4a42449" (UID: "9a08e52d-bd79-4511-ab2e-86dfb4a42449"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.884936 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a08e52d-bd79-4511-ab2e-86dfb4a42449" (UID: "9a08e52d-bd79-4511-ab2e-86dfb4a42449"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.892638 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.892666 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.892677 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.892687 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25x9r\" (UniqueName: \"kubernetes.io/projected/9a08e52d-bd79-4511-ab2e-86dfb4a42449-kube-api-access-25x9r\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.892695 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.911573 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a08e52d-bd79-4511-ab2e-86dfb4a42449" (UID: "9a08e52d-bd79-4511-ab2e-86dfb4a42449"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:17:51 crc kubenswrapper[4894]: I1210 19:17:51.995081 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a08e52d-bd79-4511-ab2e-86dfb4a42449-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:52 crc kubenswrapper[4894]: I1210 19:17:52.527646 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" event={"ID":"9a08e52d-bd79-4511-ab2e-86dfb4a42449","Type":"ContainerDied","Data":"af0b44411d47272a828aaaf41fa6cec3f568e876d28b0ada56e5aa9176578d1c"} Dec 10 19:17:52 crc kubenswrapper[4894]: I1210 19:17:52.527704 4894 scope.go:117] "RemoveContainer" containerID="42af34c411f9b80c3afb60391e4a65aa54154b78cc246261bf626674ae222948" Dec 10 19:17:52 crc kubenswrapper[4894]: I1210 19:17:52.527702 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2frqm" Dec 10 19:17:52 crc kubenswrapper[4894]: I1210 19:17:52.573518 4894 scope.go:117] "RemoveContainer" containerID="b45df4516f13da7a744ce9eca1f46dfd5286f2bf0ca611127475e99186420f8c" Dec 10 19:17:52 crc kubenswrapper[4894]: I1210 19:17:52.578033 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2frqm"] Dec 10 19:17:52 crc kubenswrapper[4894]: I1210 19:17:52.598153 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2frqm"] Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.552610 4894 generic.go:334] "Generic (PLEG): container finished" podID="67b548c7-9e9f-4c17-89c7-96df0a6670b0" containerID="bfc993891b2abfaea1dad50a4536709dd87497086cd7b8728c9859894d362fec" exitCode=0 Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.558609 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" path="/var/lib/kubelet/pods/9a08e52d-bd79-4511-ab2e-86dfb4a42449/volumes" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.559299 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"67b548c7-9e9f-4c17-89c7-96df0a6670b0","Type":"ContainerDied","Data":"bfc993891b2abfaea1dad50a4536709dd87497086cd7b8728c9859894d362fec"} Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.559332 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"67b548c7-9e9f-4c17-89c7-96df0a6670b0","Type":"ContainerDied","Data":"208c1699dee0180c01ec259acac9adbf10813b1d5afe010aab9e8347ed59237f"} Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.559344 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="208c1699dee0180c01ec259acac9adbf10813b1d5afe010aab9e8347ed59237f" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.588775 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.630914 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-combined-ca-bundle\") pod \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.630975 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-scripts\") pod \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.631066 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data\") pod \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.631157 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data-custom\") pod \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.631265 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg7tz\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-kube-api-access-qg7tz\") pod \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.631307 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-certs\") pod \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\" (UID: \"67b548c7-9e9f-4c17-89c7-96df0a6670b0\") " Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.638983 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67b548c7-9e9f-4c17-89c7-96df0a6670b0" (UID: "67b548c7-9e9f-4c17-89c7-96df0a6670b0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.640027 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-certs" (OuterVolumeSpecName: "certs") pod "67b548c7-9e9f-4c17-89c7-96df0a6670b0" (UID: "67b548c7-9e9f-4c17-89c7-96df0a6670b0"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.644046 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-kube-api-access-qg7tz" (OuterVolumeSpecName: "kube-api-access-qg7tz") pod "67b548c7-9e9f-4c17-89c7-96df0a6670b0" (UID: "67b548c7-9e9f-4c17-89c7-96df0a6670b0"). InnerVolumeSpecName "kube-api-access-qg7tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.671717 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-scripts" (OuterVolumeSpecName: "scripts") pod "67b548c7-9e9f-4c17-89c7-96df0a6670b0" (UID: "67b548c7-9e9f-4c17-89c7-96df0a6670b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.671970 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67b548c7-9e9f-4c17-89c7-96df0a6670b0" (UID: "67b548c7-9e9f-4c17-89c7-96df0a6670b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.676925 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data" (OuterVolumeSpecName: "config-data") pod "67b548c7-9e9f-4c17-89c7-96df0a6670b0" (UID: "67b548c7-9e9f-4c17-89c7-96df0a6670b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.735015 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.735054 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg7tz\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-kube-api-access-qg7tz\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.735067 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/67b548c7-9e9f-4c17-89c7-96df0a6670b0-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.735086 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.735099 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:53 crc kubenswrapper[4894]: I1210 19:17:53.735109 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b548c7-9e9f-4c17-89c7-96df0a6670b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.564534 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.590014 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-75b94f784c-g8jm2"] Dec 10 19:17:54 crc kubenswrapper[4894]: E1210 19:17:54.590537 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerName="dnsmasq-dns" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.590561 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerName="dnsmasq-dns" Dec 10 19:17:54 crc kubenswrapper[4894]: E1210 19:17:54.590580 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerName="init" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.590589 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerName="init" Dec 10 19:17:54 crc kubenswrapper[4894]: E1210 19:17:54.590604 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b548c7-9e9f-4c17-89c7-96df0a6670b0" containerName="cloudkitty-proc" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.590613 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b548c7-9e9f-4c17-89c7-96df0a6670b0" containerName="cloudkitty-proc" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.590882 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b548c7-9e9f-4c17-89c7-96df0a6670b0" containerName="cloudkitty-proc" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.590909 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a08e52d-bd79-4511-ab2e-86dfb4a42449" containerName="dnsmasq-dns" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.592292 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.596087 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.596507 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.601992 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.605054 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-75b94f784c-g8jm2"] Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.620020 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.643485 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652562 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-log-httpd\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652609 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-config-data\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652695 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-run-httpd\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652723 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rzfr\" (UniqueName: \"kubernetes.io/projected/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-kube-api-access-8rzfr\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652824 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-combined-ca-bundle\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652897 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-internal-tls-certs\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652929 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-etc-swift\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.652956 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-public-tls-certs\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.662918 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.664276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.667914 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.707305 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754432 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-combined-ca-bundle\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-internal-tls-certs\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754550 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-etc-swift\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754576 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-public-tls-certs\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754609 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-config-data\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754622 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-log-httpd\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754642 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-certs\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754668 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-scripts\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754686 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754729 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-run-httpd\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754757 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rzfr\" (UniqueName: \"kubernetes.io/projected/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-kube-api-access-8rzfr\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.754834 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wc4l\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-kube-api-access-9wc4l\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.755787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-log-httpd\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.760118 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-internal-tls-certs\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.762196 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-combined-ca-bundle\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.762277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-run-httpd\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.764346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-public-tls-certs\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.764550 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-config-data\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.764717 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-etc-swift\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.778822 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rzfr\" (UniqueName: \"kubernetes.io/projected/7c4b6f49-6469-47df-9f7c-4810b5bde4f0-kube-api-access-8rzfr\") pod \"swift-proxy-75b94f784c-g8jm2\" (UID: \"7c4b6f49-6469-47df-9f7c-4810b5bde4f0\") " pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.856406 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-certs\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.856740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-scripts\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.856890 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.857107 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.857772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wc4l\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-kube-api-access-9wc4l\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.858217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.861554 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.873871 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.881334 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wc4l\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-kube-api-access-9wc4l\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.883654 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.883662 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-certs\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.883896 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-scripts\") pod \"cloudkitty-proc-0\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.911598 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:54 crc kubenswrapper[4894]: I1210 19:17:54.996401 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:17:55 crc kubenswrapper[4894]: W1210 19:17:55.486738 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod228bcf93_2f3d_461a_a006_b5429740bf6d.slice/crio-cdecde943d49e772ab686d40503745d965153b26ea15d65028de147043455bb7 WatchSource:0}: Error finding container cdecde943d49e772ab686d40503745d965153b26ea15d65028de147043455bb7: Status 404 returned error can't find the container with id cdecde943d49e772ab686d40503745d965153b26ea15d65028de147043455bb7 Dec 10 19:17:55 crc kubenswrapper[4894]: I1210 19:17:55.487481 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:17:55 crc kubenswrapper[4894]: I1210 19:17:55.569796 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67b548c7-9e9f-4c17-89c7-96df0a6670b0" path="/var/lib/kubelet/pods/67b548c7-9e9f-4c17-89c7-96df0a6670b0/volumes" Dec 10 19:17:55 crc kubenswrapper[4894]: I1210 19:17:55.576815 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-75b94f784c-g8jm2"] Dec 10 19:17:55 crc kubenswrapper[4894]: I1210 19:17:55.578612 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"228bcf93-2f3d-461a-a006-b5429740bf6d","Type":"ContainerStarted","Data":"cdecde943d49e772ab686d40503745d965153b26ea15d65028de147043455bb7"} Dec 10 19:17:55 crc kubenswrapper[4894]: W1210 19:17:55.583054 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c4b6f49_6469_47df_9f7c_4810b5bde4f0.slice/crio-cacae4f699bb287bf18328db6e434527f7e092c788f90cca8edf081039593a95 WatchSource:0}: Error finding container cacae4f699bb287bf18328db6e434527f7e092c788f90cca8edf081039593a95: Status 404 returned error can't find the container with id cacae4f699bb287bf18328db6e434527f7e092c788f90cca8edf081039593a95 Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.592942 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-75b94f784c-g8jm2" event={"ID":"7c4b6f49-6469-47df-9f7c-4810b5bde4f0","Type":"ContainerStarted","Data":"372e752640ed338f36e6dba12ddd02367e98984ac3697f29e9fcd46f6e40df04"} Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.593596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-75b94f784c-g8jm2" event={"ID":"7c4b6f49-6469-47df-9f7c-4810b5bde4f0","Type":"ContainerStarted","Data":"f042049d03d6bc319fb9c21a5adf7fa5a9c64ca3e23f0cf650d20942784297fd"} Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.593623 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-75b94f784c-g8jm2" event={"ID":"7c4b6f49-6469-47df-9f7c-4810b5bde4f0","Type":"ContainerStarted","Data":"cacae4f699bb287bf18328db6e434527f7e092c788f90cca8edf081039593a95"} Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.593661 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.593681 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.597257 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"228bcf93-2f3d-461a-a006-b5429740bf6d","Type":"ContainerStarted","Data":"c2c50ad9b4115480c242cbd5f35d9418373a3cb54dbd904a4bf68dc7a14accad"} Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.619092 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-75b94f784c-g8jm2" podStartSLOduration=2.619071946 podStartE2EDuration="2.619071946s" podCreationTimestamp="2025-12-10 19:17:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:56.617553116 +0000 UTC m=+1359.412400918" watchObservedRunningTime="2025-12-10 19:17:56.619071946 +0000 UTC m=+1359.413919738" Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.634827 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.634808116 podStartE2EDuration="2.634808116s" podCreationTimestamp="2025-12-10 19:17:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:56.633065 +0000 UTC m=+1359.427912792" watchObservedRunningTime="2025-12-10 19:17:56.634808116 +0000 UTC m=+1359.429655918" Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.979998 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.980275 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-central-agent" containerID="cri-o://2bb580e2b2ff5d45214eac5c31af25d5aa6cae324e85cfba940e4af4840047eb" gracePeriod=30 Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.980679 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="proxy-httpd" containerID="cri-o://eecc89b8810905ee7b2dd0aac8ac03c1189da0fc52168bb64ae171030487eb37" gracePeriod=30 Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.980726 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="sg-core" containerID="cri-o://588f33e6583a95c9043727441bb20f5d0f26e5a9d72e7b6be25faa790971b429" gracePeriod=30 Dec 10 19:17:56 crc kubenswrapper[4894]: I1210 19:17:56.980756 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-notification-agent" containerID="cri-o://0a41a2abefaf26050d477881a439fbe24383960e585a45e145ac3fe34a8748c7" gracePeriod=30 Dec 10 19:17:57 crc kubenswrapper[4894]: I1210 19:17:57.613075 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerID="eecc89b8810905ee7b2dd0aac8ac03c1189da0fc52168bb64ae171030487eb37" exitCode=0 Dec 10 19:17:57 crc kubenswrapper[4894]: I1210 19:17:57.613445 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerID="588f33e6583a95c9043727441bb20f5d0f26e5a9d72e7b6be25faa790971b429" exitCode=2 Dec 10 19:17:57 crc kubenswrapper[4894]: I1210 19:17:57.613374 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerDied","Data":"eecc89b8810905ee7b2dd0aac8ac03c1189da0fc52168bb64ae171030487eb37"} Dec 10 19:17:57 crc kubenswrapper[4894]: I1210 19:17:57.614503 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerDied","Data":"588f33e6583a95c9043727441bb20f5d0f26e5a9d72e7b6be25faa790971b429"} Dec 10 19:17:58 crc kubenswrapper[4894]: I1210 19:17:58.641686 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerID="0a41a2abefaf26050d477881a439fbe24383960e585a45e145ac3fe34a8748c7" exitCode=0 Dec 10 19:17:58 crc kubenswrapper[4894]: I1210 19:17:58.654078 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerID="2bb580e2b2ff5d45214eac5c31af25d5aa6cae324e85cfba940e4af4840047eb" exitCode=0 Dec 10 19:17:58 crc kubenswrapper[4894]: I1210 19:17:58.641977 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerDied","Data":"0a41a2abefaf26050d477881a439fbe24383960e585a45e145ac3fe34a8748c7"} Dec 10 19:17:58 crc kubenswrapper[4894]: I1210 19:17:58.654143 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerDied","Data":"2bb580e2b2ff5d45214eac5c31af25d5aa6cae324e85cfba940e4af4840047eb"} Dec 10 19:18:00 crc kubenswrapper[4894]: I1210 19:18:00.458685 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:18:00 crc kubenswrapper[4894]: I1210 19:18:00.459223 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-log" containerID="cri-o://592ae354c14bfc45e128bbf3ae8dcbbc9005608ae52f5d15b85262957b401560" gracePeriod=30 Dec 10 19:18:00 crc kubenswrapper[4894]: I1210 19:18:00.459370 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-httpd" containerID="cri-o://a405b6ece97e13ee5a67ec751dd87c371f7281cf5e0cdd07d2c2400161b5d583" gracePeriod=30 Dec 10 19:18:00 crc kubenswrapper[4894]: I1210 19:18:00.674846 4894 generic.go:334] "Generic (PLEG): container finished" podID="9874ce74-c8e9-4719-802b-81a69271131a" containerID="592ae354c14bfc45e128bbf3ae8dcbbc9005608ae52f5d15b85262957b401560" exitCode=143 Dec 10 19:18:00 crc kubenswrapper[4894]: I1210 19:18:00.674891 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9874ce74-c8e9-4719-802b-81a69271131a","Type":"ContainerDied","Data":"592ae354c14bfc45e128bbf3ae8dcbbc9005608ae52f5d15b85262957b401560"} Dec 10 19:18:01 crc kubenswrapper[4894]: I1210 19:18:01.808686 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:18:01 crc kubenswrapper[4894]: I1210 19:18:01.811378 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-log" containerID="cri-o://cc88e7e330e027f3d754ca2ed27e1d86d3653e84fc0356b23537a20c616833a3" gracePeriod=30 Dec 10 19:18:01 crc kubenswrapper[4894]: I1210 19:18:01.811383 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-httpd" containerID="cri-o://ddb59b102f6178428ef1393bd729c15c7b7c711dd5fbc45982cbd4861ea55c73" gracePeriod=30 Dec 10 19:18:02 crc kubenswrapper[4894]: I1210 19:18:02.695415 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca7b0207-4a62-4299-aa2a-14244c283866" containerID="cc88e7e330e027f3d754ca2ed27e1d86d3653e84fc0356b23537a20c616833a3" exitCode=143 Dec 10 19:18:02 crc kubenswrapper[4894]: I1210 19:18:02.695498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ca7b0207-4a62-4299-aa2a-14244c283866","Type":"ContainerDied","Data":"cc88e7e330e027f3d754ca2ed27e1d86d3653e84fc0356b23537a20c616833a3"} Dec 10 19:18:03 crc kubenswrapper[4894]: I1210 19:18:03.709107 4894 generic.go:334] "Generic (PLEG): container finished" podID="9874ce74-c8e9-4719-802b-81a69271131a" containerID="a405b6ece97e13ee5a67ec751dd87c371f7281cf5e0cdd07d2c2400161b5d583" exitCode=0 Dec 10 19:18:03 crc kubenswrapper[4894]: I1210 19:18:03.709147 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9874ce74-c8e9-4719-802b-81a69271131a","Type":"ContainerDied","Data":"a405b6ece97e13ee5a67ec751dd87c371f7281cf5e0cdd07d2c2400161b5d583"} Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.133455 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.136489 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-75b94f784c-g8jm2" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.472126 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-zc6w6"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.481682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.499939 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zc6w6"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.596960 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad584c11-9601-4710-b5d4-fc6e58841d4f-operator-scripts\") pod \"nova-api-db-create-zc6w6\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.597431 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbkdj\" (UniqueName: \"kubernetes.io/projected/ad584c11-9601-4710-b5d4-fc6e58841d4f-kube-api-access-dbkdj\") pod \"nova-api-db-create-zc6w6\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.697215 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-w726g"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.698578 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.699971 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbkdj\" (UniqueName: \"kubernetes.io/projected/ad584c11-9601-4710-b5d4-fc6e58841d4f-kube-api-access-dbkdj\") pod \"nova-api-db-create-zc6w6\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.700053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad584c11-9601-4710-b5d4-fc6e58841d4f-operator-scripts\") pod \"nova-api-db-create-zc6w6\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.700657 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad584c11-9601-4710-b5d4-fc6e58841d4f-operator-scripts\") pod \"nova-api-db-create-zc6w6\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.713672 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-w726g"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.728340 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6d03-account-create-update-gcqt5"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.729682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.731465 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.733530 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbkdj\" (UniqueName: \"kubernetes.io/projected/ad584c11-9601-4710-b5d4-fc6e58841d4f-kube-api-access-dbkdj\") pod \"nova-api-db-create-zc6w6\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.745619 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6d03-account-create-update-gcqt5"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.793227 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca7b0207-4a62-4299-aa2a-14244c283866" containerID="ddb59b102f6178428ef1393bd729c15c7b7c711dd5fbc45982cbd4861ea55c73" exitCode=0 Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.794457 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ca7b0207-4a62-4299-aa2a-14244c283866","Type":"ContainerDied","Data":"ddb59b102f6178428ef1393bd729c15c7b7c711dd5fbc45982cbd4861ea55c73"} Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.803284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h54n5\" (UniqueName: \"kubernetes.io/projected/607bf25a-d450-439f-977a-2f38c41f017f-kube-api-access-h54n5\") pod \"nova-cell0-db-create-w726g\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.803363 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkzvx\" (UniqueName: \"kubernetes.io/projected/93744ff6-5842-4ea1-9d60-0859da1c5dc1-kube-api-access-gkzvx\") pod \"nova-api-6d03-account-create-update-gcqt5\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.803611 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607bf25a-d450-439f-977a-2f38c41f017f-operator-scripts\") pod \"nova-cell0-db-create-w726g\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.803707 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93744ff6-5842-4ea1-9d60-0859da1c5dc1-operator-scripts\") pod \"nova-api-6d03-account-create-update-gcqt5\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.833089 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kmkpd"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.835008 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.848318 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.877114 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kmkpd"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.887896 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-810e-account-create-update-b6bk8"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.889573 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.901022 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.904956 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkzvx\" (UniqueName: \"kubernetes.io/projected/93744ff6-5842-4ea1-9d60-0859da1c5dc1-kube-api-access-gkzvx\") pod \"nova-api-6d03-account-create-update-gcqt5\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905062 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2stm7\" (UniqueName: \"kubernetes.io/projected/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-kube-api-access-2stm7\") pod \"nova-cell1-db-create-kmkpd\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905106 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-operator-scripts\") pod \"nova-cell0-810e-account-create-update-b6bk8\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k24fc\" (UniqueName: \"kubernetes.io/projected/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-kube-api-access-k24fc\") pod \"nova-cell0-810e-account-create-update-b6bk8\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905192 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607bf25a-d450-439f-977a-2f38c41f017f-operator-scripts\") pod \"nova-cell0-db-create-w726g\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-operator-scripts\") pod \"nova-cell1-db-create-kmkpd\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905425 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93744ff6-5842-4ea1-9d60-0859da1c5dc1-operator-scripts\") pod \"nova-api-6d03-account-create-update-gcqt5\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905561 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h54n5\" (UniqueName: \"kubernetes.io/projected/607bf25a-d450-439f-977a-2f38c41f017f-kube-api-access-h54n5\") pod \"nova-cell0-db-create-w726g\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.905755 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607bf25a-d450-439f-977a-2f38c41f017f-operator-scripts\") pod \"nova-cell0-db-create-w726g\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.906821 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93744ff6-5842-4ea1-9d60-0859da1c5dc1-operator-scripts\") pod \"nova-api-6d03-account-create-update-gcqt5\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.922375 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-810e-account-create-update-b6bk8"] Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.923708 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h54n5\" (UniqueName: \"kubernetes.io/projected/607bf25a-d450-439f-977a-2f38c41f017f-kube-api-access-h54n5\") pod \"nova-cell0-db-create-w726g\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:05 crc kubenswrapper[4894]: I1210 19:18:05.946494 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkzvx\" (UniqueName: \"kubernetes.io/projected/93744ff6-5842-4ea1-9d60-0859da1c5dc1-kube-api-access-gkzvx\") pod \"nova-api-6d03-account-create-update-gcqt5\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.010886 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2stm7\" (UniqueName: \"kubernetes.io/projected/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-kube-api-access-2stm7\") pod \"nova-cell1-db-create-kmkpd\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.010941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-operator-scripts\") pod \"nova-cell0-810e-account-create-update-b6bk8\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.010994 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k24fc\" (UniqueName: \"kubernetes.io/projected/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-kube-api-access-k24fc\") pod \"nova-cell0-810e-account-create-update-b6bk8\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.011038 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-operator-scripts\") pod \"nova-cell1-db-create-kmkpd\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.011724 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-operator-scripts\") pod \"nova-cell1-db-create-kmkpd\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.012638 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-operator-scripts\") pod \"nova-cell0-810e-account-create-update-b6bk8\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.038910 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2stm7\" (UniqueName: \"kubernetes.io/projected/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-kube-api-access-2stm7\") pod \"nova-cell1-db-create-kmkpd\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.041313 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k24fc\" (UniqueName: \"kubernetes.io/projected/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-kube-api-access-k24fc\") pod \"nova-cell0-810e-account-create-update-b6bk8\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.050446 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.070738 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.085485 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.092994 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-54be-account-create-update-lxj4w"] Dec 10 19:18:06 crc kubenswrapper[4894]: E1210 19:18:06.093562 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-log" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.093581 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-log" Dec 10 19:18:06 crc kubenswrapper[4894]: E1210 19:18:06.093622 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-httpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.093631 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-httpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.093888 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-log" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.093907 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9874ce74-c8e9-4719-802b-81a69271131a" containerName="glance-httpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.094773 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.097158 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.142281 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-54be-account-create-update-lxj4w"] Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.168845 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.191475 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.221497 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-logs\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.221565 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-config-data\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.221695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-combined-ca-bundle\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.221757 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-httpd-run\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.221884 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.221918 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-scripts\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.222047 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z28xd\" (UniqueName: \"kubernetes.io/projected/9874ce74-c8e9-4719-802b-81a69271131a-kube-api-access-z28xd\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.222105 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-public-tls-certs\") pod \"9874ce74-c8e9-4719-802b-81a69271131a\" (UID: \"9874ce74-c8e9-4719-802b-81a69271131a\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.222549 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-logs" (OuterVolumeSpecName: "logs") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.222575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4qnn\" (UniqueName: \"kubernetes.io/projected/11d93ec8-1df6-456e-9043-8852cc63a87d-kube-api-access-b4qnn\") pod \"nova-cell1-54be-account-create-update-lxj4w\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.222622 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.222924 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d93ec8-1df6-456e-9043-8852cc63a87d-operator-scripts\") pod \"nova-cell1-54be-account-create-update-lxj4w\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.223168 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.223259 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9874ce74-c8e9-4719-802b-81a69271131a-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.226703 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-scripts" (OuterVolumeSpecName: "scripts") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.228282 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.231981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9874ce74-c8e9-4719-802b-81a69271131a-kube-api-access-z28xd" (OuterVolumeSpecName: "kube-api-access-z28xd") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "kube-api-access-z28xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.265844 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5" (OuterVolumeSpecName: "glance") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.267134 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.299807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.324806 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-sg-core-conf-yaml\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325205 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-run-httpd\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325596 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-scripts\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-config-data\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325663 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bcwl\" (UniqueName: \"kubernetes.io/projected/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-kube-api-access-8bcwl\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325654 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325680 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-combined-ca-bundle\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325828 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-log-httpd\") pod \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\" (UID: \"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.325996 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-config-data" (OuterVolumeSpecName: "config-data") pod "9874ce74-c8e9-4719-802b-81a69271131a" (UID: "9874ce74-c8e9-4719-802b-81a69271131a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.326470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d93ec8-1df6-456e-9043-8852cc63a87d-operator-scripts\") pod \"nova-cell1-54be-account-create-update-lxj4w\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.326722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4qnn\" (UniqueName: \"kubernetes.io/projected/11d93ec8-1df6-456e-9043-8852cc63a87d-kube-api-access-b4qnn\") pod \"nova-cell1-54be-account-create-update-lxj4w\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.326940 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z28xd\" (UniqueName: \"kubernetes.io/projected/9874ce74-c8e9-4719-802b-81a69271131a-kube-api-access-z28xd\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327141 4894 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327153 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327162 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327170 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327216 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") on node \"crc\" " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327395 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9874ce74-c8e9-4719-802b-81a69271131a-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327398 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d93ec8-1df6-456e-9043-8852cc63a87d-operator-scripts\") pod \"nova-cell1-54be-account-create-update-lxj4w\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.327295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.330481 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-scripts" (OuterVolumeSpecName: "scripts") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.332108 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-kube-api-access-8bcwl" (OuterVolumeSpecName: "kube-api-access-8bcwl") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "kube-api-access-8bcwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.350129 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4qnn\" (UniqueName: \"kubernetes.io/projected/11d93ec8-1df6-456e-9043-8852cc63a87d-kube-api-access-b4qnn\") pod \"nova-cell1-54be-account-create-update-lxj4w\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.360974 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.361145 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5") on node "crc" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.438454 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.438518 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.438528 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bcwl\" (UniqueName: \"kubernetes.io/projected/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-kube-api-access-8bcwl\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.438538 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.462530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.473349 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.533632 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.541338 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.541364 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.626102 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-config-data" (OuterVolumeSpecName: "config-data") pod "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" (UID: "7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.642187 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.675714 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.845624 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-combined-ca-bundle\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846030 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-logs\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846088 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsdv5\" (UniqueName: \"kubernetes.io/projected/ca7b0207-4a62-4299-aa2a-14244c283866-kube-api-access-qsdv5\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846116 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-internal-tls-certs\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846162 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-scripts\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846269 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-config-data\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846360 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-httpd-run\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.846550 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"ca7b0207-4a62-4299-aa2a-14244c283866\" (UID: \"ca7b0207-4a62-4299-aa2a-14244c283866\") " Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.850571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-logs" (OuterVolumeSpecName: "logs") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.854754 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-scripts" (OuterVolumeSpecName: "scripts") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.863549 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.866118 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7b0207-4a62-4299-aa2a-14244c283866-kube-api-access-qsdv5" (OuterVolumeSpecName: "kube-api-access-qsdv5") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "kube-api-access-qsdv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.867425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3d4126ef-dbcf-4614-a1fe-37c7c0be7223","Type":"ContainerStarted","Data":"2444bfa00d56738aad3a8143720a545bb1c7154d3d62c2f04d47727356c229b8"} Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.939925 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zc6w6"] Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.940519 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4","Type":"ContainerDied","Data":"b9b0b9a6a3c5d0ae70b9c9239f643cf5b5b757048cca3dbb097674f6bd518ed4"} Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.940571 4894 scope.go:117] "RemoveContainer" containerID="eecc89b8810905ee7b2dd0aac8ac03c1189da0fc52168bb64ae171030487eb37" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.940719 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.954556 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsdv5\" (UniqueName: \"kubernetes.io/projected/ca7b0207-4a62-4299-aa2a-14244c283866-kube-api-access-qsdv5\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.966464 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.966542 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:06 crc kubenswrapper[4894]: I1210 19:18:06.966611 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7b0207-4a62-4299-aa2a-14244c283866-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:06.999307 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.637247887 podStartE2EDuration="16.999289298s" podCreationTimestamp="2025-12-10 19:17:50 +0000 UTC" firstStartedPulling="2025-12-10 19:17:51.203065755 +0000 UTC m=+1353.997913547" lastFinishedPulling="2025-12-10 19:18:05.565107166 +0000 UTC m=+1368.359954958" observedRunningTime="2025-12-10 19:18:06.901066238 +0000 UTC m=+1369.695914030" watchObservedRunningTime="2025-12-10 19:18:06.999289298 +0000 UTC m=+1369.794137090" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.028686 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9874ce74-c8e9-4719-802b-81a69271131a","Type":"ContainerDied","Data":"ad8514fd3146c036a58e953aaf9e513597474c429a6b20ca325e87aff4f4bca0"} Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.028944 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.045624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.068571 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.073717 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ca7b0207-4a62-4299-aa2a-14244c283866","Type":"ContainerDied","Data":"d68d176036272d46ccf0d6f0e4b06085fdc4b61e9b215dbf65fee0e64032398e"} Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.073803 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.089978 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-w726g"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.092343 4894 scope.go:117] "RemoveContainer" containerID="588f33e6583a95c9043727441bb20f5d0f26e5a9d72e7b6be25faa790971b429" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.100513 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.123424 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.156081 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-config-data" (OuterVolumeSpecName: "config-data") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.162543 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4" (OuterVolumeSpecName: "glance") pod "ca7b0207-4a62-4299-aa2a-14244c283866" (UID: "ca7b0207-4a62-4299-aa2a-14244c283866"). InnerVolumeSpecName "pvc-102f2a85-63ab-4ac5-99ed-6047367290b4". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.168675 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.170043 4894 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.170063 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7b0207-4a62-4299-aa2a-14244c283866-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.170062 4894 scope.go:117] "RemoveContainer" containerID="0a41a2abefaf26050d477881a439fbe24383960e585a45e145ac3fe34a8748c7" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.170094 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") on node \"crc\" " Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.179315 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: E1210 19:18:07.180143 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-httpd" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.180214 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-httpd" Dec 10 19:18:07 crc kubenswrapper[4894]: E1210 19:18:07.180277 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="sg-core" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.180353 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="sg-core" Dec 10 19:18:07 crc kubenswrapper[4894]: E1210 19:18:07.180422 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-central-agent" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.180538 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-central-agent" Dec 10 19:18:07 crc kubenswrapper[4894]: E1210 19:18:07.180595 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-log" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.180676 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-log" Dec 10 19:18:07 crc kubenswrapper[4894]: E1210 19:18:07.180728 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-notification-agent" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.180787 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-notification-agent" Dec 10 19:18:07 crc kubenswrapper[4894]: E1210 19:18:07.180872 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="proxy-httpd" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.180927 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="proxy-httpd" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.181147 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-notification-agent" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.181223 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-httpd" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.181286 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" containerName="glance-log" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.181350 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="ceilometer-central-agent" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.181448 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="sg-core" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.181513 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" containerName="proxy-httpd" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.185076 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.188393 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.188587 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.196931 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.222223 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.222377 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-102f2a85-63ab-4ac5-99ed-6047367290b4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4") on node "crc" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.257453 4894 scope.go:117] "RemoveContainer" containerID="2bb580e2b2ff5d45214eac5c31af25d5aa6cae324e85cfba940e4af4840047eb" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.264970 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.285452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-run-httpd\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.285697 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-log-httpd\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.285891 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.286112 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-scripts\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.286135 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-config-data\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.286244 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.298694 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.325379 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.331397 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.336588 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.336912 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.337266 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.345330 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6d03-account-create-update-gcqt5"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.388645 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.388916 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.389048 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7x4m\" (UniqueName: \"kubernetes.io/projected/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-kube-api-access-b7x4m\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.389146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-scripts\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.389223 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-config-data\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.389359 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-run-httpd\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.389466 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-log-httpd\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.390068 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-log-httpd\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.393947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.402194 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-scripts\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.407153 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-run-httpd\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.409994 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-config-data\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.466252 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-810e-account-create-update-b6bk8"] Dec 10 19:18:07 crc kubenswrapper[4894]: W1210 19:18:07.475147 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5ba7db9_ba2c_4cb7_9edc_2ed985264548.slice/crio-1ce4914403d10cc551a1a2339e973e89b59d829d6d2b0c039543ae93b81d9046 WatchSource:0}: Error finding container 1ce4914403d10cc551a1a2339e973e89b59d829d6d2b0c039543ae93b81d9046: Status 404 returned error can't find the container with id 1ce4914403d10cc551a1a2339e973e89b59d829d6d2b0c039543ae93b81d9046 Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491511 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491554 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491581 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7x4m\" (UniqueName: \"kubernetes.io/projected/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-kube-api-access-b7x4m\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491644 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-config-data\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491666 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-logs\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491696 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwzpz\" (UniqueName: \"kubernetes.io/projected/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-kube-api-access-qwzpz\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491771 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-scripts\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.491811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.501125 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kmkpd"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.506316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.517931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7x4m\" (UniqueName: \"kubernetes.io/projected/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-kube-api-access-b7x4m\") pod \"ceilometer-0\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.562218 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4" path="/var/lib/kubelet/pods/7fb5bb50-8176-4ff7-aea8-4c58bd0b07b4/volumes" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.563453 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9874ce74-c8e9-4719-802b-81a69271131a" path="/var/lib/kubelet/pods/9874ce74-c8e9-4719-802b-81a69271131a/volumes" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.583888 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-54be-account-create-update-lxj4w"] Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594287 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-config-data\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594353 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-logs\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwzpz\" (UniqueName: \"kubernetes.io/projected/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-kube-api-access-qwzpz\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594454 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.594526 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-scripts\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.595762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.596729 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-logs\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.599682 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-scripts\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.603810 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-config-data\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.607461 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.608481 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.618505 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwzpz\" (UniqueName: \"kubernetes.io/projected/e485201c-3c59-4daa-ba4d-f7cb8e91c3ec-kube-api-access-qwzpz\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.621218 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.621259 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6b95216d4ff1b4536f5681f700c7b5d7fd226b07d8f87d718c34fcdf28bde918/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.627502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.642075 4894 scope.go:117] "RemoveContainer" containerID="a405b6ece97e13ee5a67ec751dd87c371f7281cf5e0cdd07d2c2400161b5d583" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.673720 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d9c74b8a-cf8d-41bd-82d9-889ce5159ea5\") pod \"glance-default-external-api-0\" (UID: \"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec\") " pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.734709 4894 scope.go:117] "RemoveContainer" containerID="592ae354c14bfc45e128bbf3ae8dcbbc9005608ae52f5d15b85262957b401560" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.743460 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.831410 4894 scope.go:117] "RemoveContainer" containerID="ddb59b102f6178428ef1393bd729c15c7b7c711dd5fbc45982cbd4861ea55c73" Dec 10 19:18:07 crc kubenswrapper[4894]: I1210 19:18:07.885783 4894 scope.go:117] "RemoveContainer" containerID="cc88e7e330e027f3d754ca2ed27e1d86d3653e84fc0356b23537a20c616833a3" Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.095282 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" event={"ID":"11d93ec8-1df6-456e-9043-8852cc63a87d","Type":"ContainerStarted","Data":"e2e992155b59dcd2f5f248a8d687f2d3d01529b6648fcb5327966142cbb345f3"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.095339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" event={"ID":"11d93ec8-1df6-456e-9043-8852cc63a87d","Type":"ContainerStarted","Data":"763157e1430f0073089100bb5e91afe951cb328cb5f27a9bf96f43800a458330"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.110157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kmkpd" event={"ID":"33f598ed-63ab-4a69-acb3-33fe9b9fabd8","Type":"ContainerStarted","Data":"4081dd318e2ae9bc713fa54e69070b48df44710fcc6d14514adcba17887ff2a1"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.113509 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6d03-account-create-update-gcqt5" event={"ID":"93744ff6-5842-4ea1-9d60-0859da1c5dc1","Type":"ContainerStarted","Data":"2bef1b253b1b5483e4fd58ba310bf18366e3ff5174d13d377696dc421db3f0ec"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.113559 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6d03-account-create-update-gcqt5" event={"ID":"93744ff6-5842-4ea1-9d60-0859da1c5dc1","Type":"ContainerStarted","Data":"f39234d6016d3a6e08d5e94ba08cfe9cf972897023b50de1817853234b9ac2ea"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.126016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zc6w6" event={"ID":"ad584c11-9601-4710-b5d4-fc6e58841d4f","Type":"ContainerStarted","Data":"e6ea07194db43ce02d28708a77667b104f90088b218e1a94ff0e0526c7a8001f"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.126042 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zc6w6" event={"ID":"ad584c11-9601-4710-b5d4-fc6e58841d4f","Type":"ContainerStarted","Data":"6c8630e6815fcd80d7e1713cfbb886effdabcca0a65f5dd99beb6d7b9befd101"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.134288 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" podStartSLOduration=2.134267017 podStartE2EDuration="2.134267017s" podCreationTimestamp="2025-12-10 19:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:08.119187519 +0000 UTC m=+1370.914035321" watchObservedRunningTime="2025-12-10 19:18:08.134267017 +0000 UTC m=+1370.929114809" Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.159539 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" event={"ID":"e5ba7db9-ba2c-4cb7-9edc-2ed985264548","Type":"ContainerStarted","Data":"b5da647cf904e8ce6e2a76e4dfe990f116ce27b895b253af3d8122534975e540"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.159584 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" event={"ID":"e5ba7db9-ba2c-4cb7-9edc-2ed985264548","Type":"ContainerStarted","Data":"1ce4914403d10cc551a1a2339e973e89b59d829d6d2b0c039543ae93b81d9046"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.170929 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w726g" event={"ID":"607bf25a-d450-439f-977a-2f38c41f017f","Type":"ContainerStarted","Data":"519dcac4fb39eca4e80f88d59f0f0cda59ae5ba20f6def3e8c68acc7844d35a8"} Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.178374 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-6d03-account-create-update-gcqt5" podStartSLOduration=3.178356351 podStartE2EDuration="3.178356351s" podCreationTimestamp="2025-12-10 19:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:08.132325605 +0000 UTC m=+1370.927173407" watchObservedRunningTime="2025-12-10 19:18:08.178356351 +0000 UTC m=+1370.973204143" Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.222881 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-zc6w6" podStartSLOduration=3.2228377249999998 podStartE2EDuration="3.222837725s" podCreationTimestamp="2025-12-10 19:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:08.151397191 +0000 UTC m=+1370.946244983" watchObservedRunningTime="2025-12-10 19:18:08.222837725 +0000 UTC m=+1371.017685537" Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.241624 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.250555 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" podStartSLOduration=3.250530975 podStartE2EDuration="3.250530975s" podCreationTimestamp="2025-12-10 19:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:08.183588852 +0000 UTC m=+1370.978436644" watchObservedRunningTime="2025-12-10 19:18:08.250530975 +0000 UTC m=+1371.045378767" Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.261256 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-w726g" podStartSLOduration=3.261230945 podStartE2EDuration="3.261230945s" podCreationTimestamp="2025-12-10 19:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:08.196748979 +0000 UTC m=+1370.991596771" watchObservedRunningTime="2025-12-10 19:18:08.261230945 +0000 UTC m=+1371.056078737" Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.453566 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:18:08 crc kubenswrapper[4894]: I1210 19:18:08.631278 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.186498 4894 generic.go:334] "Generic (PLEG): container finished" podID="93744ff6-5842-4ea1-9d60-0859da1c5dc1" containerID="2bef1b253b1b5483e4fd58ba310bf18366e3ff5174d13d377696dc421db3f0ec" exitCode=0 Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.186826 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6d03-account-create-update-gcqt5" event={"ID":"93744ff6-5842-4ea1-9d60-0859da1c5dc1","Type":"ContainerDied","Data":"2bef1b253b1b5483e4fd58ba310bf18366e3ff5174d13d377696dc421db3f0ec"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.197620 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5ba7db9-ba2c-4cb7-9edc-2ed985264548" containerID="b5da647cf904e8ce6e2a76e4dfe990f116ce27b895b253af3d8122534975e540" exitCode=0 Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.197710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" event={"ID":"e5ba7db9-ba2c-4cb7-9edc-2ed985264548","Type":"ContainerDied","Data":"b5da647cf904e8ce6e2a76e4dfe990f116ce27b895b253af3d8122534975e540"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.208066 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec","Type":"ContainerStarted","Data":"0d64a4d096d8c733df71c96d6e6978bb291b71c765d0bccc48208e5ffabad268"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.216056 4894 generic.go:334] "Generic (PLEG): container finished" podID="607bf25a-d450-439f-977a-2f38c41f017f" containerID="e8b544c7dd401ce2ba5607d298b1d009e44cd9de07eed07c0ce0788cc29530fd" exitCode=0 Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.216125 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w726g" event={"ID":"607bf25a-d450-439f-977a-2f38c41f017f","Type":"ContainerDied","Data":"e8b544c7dd401ce2ba5607d298b1d009e44cd9de07eed07c0ce0788cc29530fd"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.218678 4894 generic.go:334] "Generic (PLEG): container finished" podID="ad584c11-9601-4710-b5d4-fc6e58841d4f" containerID="e6ea07194db43ce02d28708a77667b104f90088b218e1a94ff0e0526c7a8001f" exitCode=0 Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.218749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zc6w6" event={"ID":"ad584c11-9601-4710-b5d4-fc6e58841d4f","Type":"ContainerDied","Data":"e6ea07194db43ce02d28708a77667b104f90088b218e1a94ff0e0526c7a8001f"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.224925 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerStarted","Data":"732b6fa634b14916afc415876f659e3ae71bf85a731b0d0c10b71e7856b55b3b"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.236324 4894 generic.go:334] "Generic (PLEG): container finished" podID="11d93ec8-1df6-456e-9043-8852cc63a87d" containerID="e2e992155b59dcd2f5f248a8d687f2d3d01529b6648fcb5327966142cbb345f3" exitCode=0 Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.236397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" event={"ID":"11d93ec8-1df6-456e-9043-8852cc63a87d","Type":"ContainerDied","Data":"e2e992155b59dcd2f5f248a8d687f2d3d01529b6648fcb5327966142cbb345f3"} Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.243020 4894 generic.go:334] "Generic (PLEG): container finished" podID="33f598ed-63ab-4a69-acb3-33fe9b9fabd8" containerID="d3ccec34a613b7f6d82204c1d1f81ef7f348e9e33fee501e976454437c92deff" exitCode=0 Dec 10 19:18:09 crc kubenswrapper[4894]: I1210 19:18:09.243079 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kmkpd" event={"ID":"33f598ed-63ab-4a69-acb3-33fe9b9fabd8","Type":"ContainerDied","Data":"d3ccec34a613b7f6d82204c1d1f81ef7f348e9e33fee501e976454437c92deff"} Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.255356 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec","Type":"ContainerStarted","Data":"a91b81d04545f8400ad21a9d52f1a75b55e9cb1561ce4e0e1b6e5ead4c80c445"} Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.255777 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e485201c-3c59-4daa-ba4d-f7cb8e91c3ec","Type":"ContainerStarted","Data":"c785a027abf714ba01f7d96ed51dbe7d9c8ca046faf9e9e7f62d50499ab90d28"} Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.259254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerStarted","Data":"81c920761368c090c0dafbc881b940ffc060c03579a7436c924bd2bb9cee1ee7"} Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.259288 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerStarted","Data":"f620f337eb625e5206af5aaf50840b89eb91103c654cfeeab5dfd732b56c32d8"} Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.282916 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.282898192 podStartE2EDuration="3.282898192s" podCreationTimestamp="2025-12-10 19:18:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:10.274819383 +0000 UTC m=+1373.069667195" watchObservedRunningTime="2025-12-10 19:18:10.282898192 +0000 UTC m=+1373.077745984" Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.867021 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.988639 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkzvx\" (UniqueName: \"kubernetes.io/projected/93744ff6-5842-4ea1-9d60-0859da1c5dc1-kube-api-access-gkzvx\") pod \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.989313 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93744ff6-5842-4ea1-9d60-0859da1c5dc1-operator-scripts\") pod \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\" (UID: \"93744ff6-5842-4ea1-9d60-0859da1c5dc1\") " Dec 10 19:18:10 crc kubenswrapper[4894]: I1210 19:18:10.990870 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93744ff6-5842-4ea1-9d60-0859da1c5dc1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "93744ff6-5842-4ea1-9d60-0859da1c5dc1" (UID: "93744ff6-5842-4ea1-9d60-0859da1c5dc1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.007974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93744ff6-5842-4ea1-9d60-0859da1c5dc1-kube-api-access-gkzvx" (OuterVolumeSpecName: "kube-api-access-gkzvx") pod "93744ff6-5842-4ea1-9d60-0859da1c5dc1" (UID: "93744ff6-5842-4ea1-9d60-0859da1c5dc1"). InnerVolumeSpecName "kube-api-access-gkzvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.092382 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkzvx\" (UniqueName: \"kubernetes.io/projected/93744ff6-5842-4ea1-9d60-0859da1c5dc1-kube-api-access-gkzvx\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.092416 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93744ff6-5842-4ea1-9d60-0859da1c5dc1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.194000 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.209454 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.231659 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.243573 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.248602 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.284098 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zc6w6" event={"ID":"ad584c11-9601-4710-b5d4-fc6e58841d4f","Type":"ContainerDied","Data":"6c8630e6815fcd80d7e1713cfbb886effdabcca0a65f5dd99beb6d7b9befd101"} Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.284136 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c8630e6815fcd80d7e1713cfbb886effdabcca0a65f5dd99beb6d7b9befd101" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.284194 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zc6w6" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.294653 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295087 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-operator-scripts\") pod \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295156 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54be-account-create-update-lxj4w" event={"ID":"11d93ec8-1df6-456e-9043-8852cc63a87d","Type":"ContainerDied","Data":"763157e1430f0073089100bb5e91afe951cb328cb5f27a9bf96f43800a458330"} Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295191 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h54n5\" (UniqueName: \"kubernetes.io/projected/607bf25a-d450-439f-977a-2f38c41f017f-kube-api-access-h54n5\") pod \"607bf25a-d450-439f-977a-2f38c41f017f\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295270 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d93ec8-1df6-456e-9043-8852cc63a87d-operator-scripts\") pod \"11d93ec8-1df6-456e-9043-8852cc63a87d\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295306 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4qnn\" (UniqueName: \"kubernetes.io/projected/11d93ec8-1df6-456e-9043-8852cc63a87d-kube-api-access-b4qnn\") pod \"11d93ec8-1df6-456e-9043-8852cc63a87d\" (UID: \"11d93ec8-1df6-456e-9043-8852cc63a87d\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295352 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad584c11-9601-4710-b5d4-fc6e58841d4f-operator-scripts\") pod \"ad584c11-9601-4710-b5d4-fc6e58841d4f\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295405 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607bf25a-d450-439f-977a-2f38c41f017f-operator-scripts\") pod \"607bf25a-d450-439f-977a-2f38c41f017f\" (UID: \"607bf25a-d450-439f-977a-2f38c41f017f\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295511 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k24fc\" (UniqueName: \"kubernetes.io/projected/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-kube-api-access-k24fc\") pod \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\" (UID: \"e5ba7db9-ba2c-4cb7-9edc-2ed985264548\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295560 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-operator-scripts\") pod \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbkdj\" (UniqueName: \"kubernetes.io/projected/ad584c11-9601-4710-b5d4-fc6e58841d4f-kube-api-access-dbkdj\") pod \"ad584c11-9601-4710-b5d4-fc6e58841d4f\" (UID: \"ad584c11-9601-4710-b5d4-fc6e58841d4f\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295672 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2stm7\" (UniqueName: \"kubernetes.io/projected/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-kube-api-access-2stm7\") pod \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\" (UID: \"33f598ed-63ab-4a69-acb3-33fe9b9fabd8\") " Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.298008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad584c11-9601-4710-b5d4-fc6e58841d4f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad584c11-9601-4710-b5d4-fc6e58841d4f" (UID: "ad584c11-9601-4710-b5d4-fc6e58841d4f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.298593 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5ba7db9-ba2c-4cb7-9edc-2ed985264548" (UID: "e5ba7db9-ba2c-4cb7-9edc-2ed985264548"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.299170 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d93ec8-1df6-456e-9043-8852cc63a87d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11d93ec8-1df6-456e-9043-8852cc63a87d" (UID: "11d93ec8-1df6-456e-9043-8852cc63a87d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.295196 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="763157e1430f0073089100bb5e91afe951cb328cb5f27a9bf96f43800a458330" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.300530 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.300556 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d93ec8-1df6-456e-9043-8852cc63a87d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.300566 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad584c11-9601-4710-b5d4-fc6e58841d4f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.300862 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607bf25a-d450-439f-977a-2f38c41f017f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "607bf25a-d450-439f-977a-2f38c41f017f" (UID: "607bf25a-d450-439f-977a-2f38c41f017f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.301526 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/607bf25a-d450-439f-977a-2f38c41f017f-kube-api-access-h54n5" (OuterVolumeSpecName: "kube-api-access-h54n5") pod "607bf25a-d450-439f-977a-2f38c41f017f" (UID: "607bf25a-d450-439f-977a-2f38c41f017f"). InnerVolumeSpecName "kube-api-access-h54n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.302014 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33f598ed-63ab-4a69-acb3-33fe9b9fabd8" (UID: "33f598ed-63ab-4a69-acb3-33fe9b9fabd8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.304440 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-kube-api-access-2stm7" (OuterVolumeSpecName: "kube-api-access-2stm7") pod "33f598ed-63ab-4a69-acb3-33fe9b9fabd8" (UID: "33f598ed-63ab-4a69-acb3-33fe9b9fabd8"). InnerVolumeSpecName "kube-api-access-2stm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.305225 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11d93ec8-1df6-456e-9043-8852cc63a87d-kube-api-access-b4qnn" (OuterVolumeSpecName: "kube-api-access-b4qnn") pod "11d93ec8-1df6-456e-9043-8852cc63a87d" (UID: "11d93ec8-1df6-456e-9043-8852cc63a87d"). InnerVolumeSpecName "kube-api-access-b4qnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.308085 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-kube-api-access-k24fc" (OuterVolumeSpecName: "kube-api-access-k24fc") pod "e5ba7db9-ba2c-4cb7-9edc-2ed985264548" (UID: "e5ba7db9-ba2c-4cb7-9edc-2ed985264548"). InnerVolumeSpecName "kube-api-access-k24fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.309944 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad584c11-9601-4710-b5d4-fc6e58841d4f-kube-api-access-dbkdj" (OuterVolumeSpecName: "kube-api-access-dbkdj") pod "ad584c11-9601-4710-b5d4-fc6e58841d4f" (UID: "ad584c11-9601-4710-b5d4-fc6e58841d4f"). InnerVolumeSpecName "kube-api-access-dbkdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.316193 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kmkpd" event={"ID":"33f598ed-63ab-4a69-acb3-33fe9b9fabd8","Type":"ContainerDied","Data":"4081dd318e2ae9bc713fa54e69070b48df44710fcc6d14514adcba17887ff2a1"} Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.316234 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4081dd318e2ae9bc713fa54e69070b48df44710fcc6d14514adcba17887ff2a1" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.316273 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kmkpd" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.318138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6d03-account-create-update-gcqt5" event={"ID":"93744ff6-5842-4ea1-9d60-0859da1c5dc1","Type":"ContainerDied","Data":"f39234d6016d3a6e08d5e94ba08cfe9cf972897023b50de1817853234b9ac2ea"} Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.318167 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f39234d6016d3a6e08d5e94ba08cfe9cf972897023b50de1817853234b9ac2ea" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.318190 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6d03-account-create-update-gcqt5" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.319557 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.335406 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-810e-account-create-update-b6bk8" event={"ID":"e5ba7db9-ba2c-4cb7-9edc-2ed985264548","Type":"ContainerDied","Data":"1ce4914403d10cc551a1a2339e973e89b59d829d6d2b0c039543ae93b81d9046"} Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.335681 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce4914403d10cc551a1a2339e973e89b59d829d6d2b0c039543ae93b81d9046" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.338632 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w726g" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.341215 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w726g" event={"ID":"607bf25a-d450-439f-977a-2f38c41f017f","Type":"ContainerDied","Data":"519dcac4fb39eca4e80f88d59f0f0cda59ae5ba20f6def3e8c68acc7844d35a8"} Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.341411 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="519dcac4fb39eca4e80f88d59f0f0cda59ae5ba20f6def3e8c68acc7844d35a8" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404613 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h54n5\" (UniqueName: \"kubernetes.io/projected/607bf25a-d450-439f-977a-2f38c41f017f-kube-api-access-h54n5\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404640 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4qnn\" (UniqueName: \"kubernetes.io/projected/11d93ec8-1df6-456e-9043-8852cc63a87d-kube-api-access-b4qnn\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404649 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607bf25a-d450-439f-977a-2f38c41f017f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404658 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k24fc\" (UniqueName: \"kubernetes.io/projected/e5ba7db9-ba2c-4cb7-9edc-2ed985264548-kube-api-access-k24fc\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404667 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404677 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbkdj\" (UniqueName: \"kubernetes.io/projected/ad584c11-9601-4710-b5d4-fc6e58841d4f-kube-api-access-dbkdj\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:11 crc kubenswrapper[4894]: I1210 19:18:11.404686 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2stm7\" (UniqueName: \"kubernetes.io/projected/33f598ed-63ab-4a69-acb3-33fe9b9fabd8-kube-api-access-2stm7\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:12 crc kubenswrapper[4894]: I1210 19:18:12.350702 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerStarted","Data":"6e7e9447eb5a49a984363a368a3ffdf30e7216d7a0c2b036d48b7df7094d5c4b"} Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.389315 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerStarted","Data":"4062b40e399e0ec3e7808e5263fec0bf8ac81cf9762ec345e153da7a5140f5d0"} Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.389708 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-central-agent" containerID="cri-o://f620f337eb625e5206af5aaf50840b89eb91103c654cfeeab5dfd732b56c32d8" gracePeriod=30 Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.390011 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.390275 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="proxy-httpd" containerID="cri-o://4062b40e399e0ec3e7808e5263fec0bf8ac81cf9762ec345e153da7a5140f5d0" gracePeriod=30 Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.390315 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="sg-core" containerID="cri-o://6e7e9447eb5a49a984363a368a3ffdf30e7216d7a0c2b036d48b7df7094d5c4b" gracePeriod=30 Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.390347 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-notification-agent" containerID="cri-o://81c920761368c090c0dafbc881b940ffc060c03579a7436c924bd2bb9cee1ee7" gracePeriod=30 Dec 10 19:18:13 crc kubenswrapper[4894]: I1210 19:18:13.427526 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.506496676 podStartE2EDuration="6.427503063s" podCreationTimestamp="2025-12-10 19:18:07 +0000 UTC" firstStartedPulling="2025-12-10 19:18:08.196154953 +0000 UTC m=+1370.991002735" lastFinishedPulling="2025-12-10 19:18:13.11716132 +0000 UTC m=+1375.912009122" observedRunningTime="2025-12-10 19:18:13.415333544 +0000 UTC m=+1376.210181356" watchObservedRunningTime="2025-12-10 19:18:13.427503063 +0000 UTC m=+1376.222350855" Dec 10 19:18:14 crc kubenswrapper[4894]: I1210 19:18:14.405899 4894 generic.go:334] "Generic (PLEG): container finished" podID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerID="6e7e9447eb5a49a984363a368a3ffdf30e7216d7a0c2b036d48b7df7094d5c4b" exitCode=2 Dec 10 19:18:14 crc kubenswrapper[4894]: I1210 19:18:14.406255 4894 generic.go:334] "Generic (PLEG): container finished" podID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerID="81c920761368c090c0dafbc881b940ffc060c03579a7436c924bd2bb9cee1ee7" exitCode=0 Dec 10 19:18:14 crc kubenswrapper[4894]: I1210 19:18:14.405948 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerDied","Data":"6e7e9447eb5a49a984363a368a3ffdf30e7216d7a0c2b036d48b7df7094d5c4b"} Dec 10 19:18:14 crc kubenswrapper[4894]: I1210 19:18:14.406309 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerDied","Data":"81c920761368c090c0dafbc881b940ffc060c03579a7436c924bd2bb9cee1ee7"} Dec 10 19:18:14 crc kubenswrapper[4894]: I1210 19:18:14.406356 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerDied","Data":"f620f337eb625e5206af5aaf50840b89eb91103c654cfeeab5dfd732b56c32d8"} Dec 10 19:18:14 crc kubenswrapper[4894]: I1210 19:18:14.406264 4894 generic.go:334] "Generic (PLEG): container finished" podID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerID="f620f337eb625e5206af5aaf50840b89eb91103c654cfeeab5dfd732b56c32d8" exitCode=0 Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.240438 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k84gv"] Dec 10 19:18:16 crc kubenswrapper[4894]: E1210 19:18:16.241216 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93744ff6-5842-4ea1-9d60-0859da1c5dc1" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241232 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="93744ff6-5842-4ea1-9d60-0859da1c5dc1" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: E1210 19:18:16.241250 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ba7db9-ba2c-4cb7-9edc-2ed985264548" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241257 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ba7db9-ba2c-4cb7-9edc-2ed985264548" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: E1210 19:18:16.241274 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad584c11-9601-4710-b5d4-fc6e58841d4f" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241282 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad584c11-9601-4710-b5d4-fc6e58841d4f" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: E1210 19:18:16.241302 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11d93ec8-1df6-456e-9043-8852cc63a87d" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241310 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="11d93ec8-1df6-456e-9043-8852cc63a87d" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: E1210 19:18:16.241333 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f598ed-63ab-4a69-acb3-33fe9b9fabd8" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241339 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f598ed-63ab-4a69-acb3-33fe9b9fabd8" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: E1210 19:18:16.241356 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607bf25a-d450-439f-977a-2f38c41f017f" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241363 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="607bf25a-d450-439f-977a-2f38c41f017f" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241566 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f598ed-63ab-4a69-acb3-33fe9b9fabd8" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241581 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad584c11-9601-4710-b5d4-fc6e58841d4f" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241595 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="607bf25a-d450-439f-977a-2f38c41f017f" containerName="mariadb-database-create" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241615 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="93744ff6-5842-4ea1-9d60-0859da1c5dc1" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241629 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="11d93ec8-1df6-456e-9043-8852cc63a87d" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.241649 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ba7db9-ba2c-4cb7-9edc-2ed985264548" containerName="mariadb-account-create-update" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.242576 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.245314 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.245836 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.246095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2479m" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.260524 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k84gv"] Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.301322 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-scripts\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.301452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.301474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-config-data\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.301500 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjc4v\" (UniqueName: \"kubernetes.io/projected/53a93527-c2bb-4968-862c-02c60d6fe93c-kube-api-access-wjc4v\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.403341 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-scripts\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.403787 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.403905 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-config-data\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.404027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjc4v\" (UniqueName: \"kubernetes.io/projected/53a93527-c2bb-4968-862c-02c60d6fe93c-kube-api-access-wjc4v\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.409275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.409655 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-scripts\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.423627 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-config-data\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.427863 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjc4v\" (UniqueName: \"kubernetes.io/projected/53a93527-c2bb-4968-862c-02c60d6fe93c-kube-api-access-wjc4v\") pod \"nova-cell0-conductor-db-sync-k84gv\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:16 crc kubenswrapper[4894]: I1210 19:18:16.559157 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:17 crc kubenswrapper[4894]: I1210 19:18:17.027448 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k84gv"] Dec 10 19:18:17 crc kubenswrapper[4894]: W1210 19:18:17.030078 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53a93527_c2bb_4968_862c_02c60d6fe93c.slice/crio-3b163d8f9376d580288539c760968d9521db0b275dfbf8b67eb029ae70713ad3 WatchSource:0}: Error finding container 3b163d8f9376d580288539c760968d9521db0b275dfbf8b67eb029ae70713ad3: Status 404 returned error can't find the container with id 3b163d8f9376d580288539c760968d9521db0b275dfbf8b67eb029ae70713ad3 Dec 10 19:18:17 crc kubenswrapper[4894]: I1210 19:18:17.450551 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k84gv" event={"ID":"53a93527-c2bb-4968-862c-02c60d6fe93c","Type":"ContainerStarted","Data":"3b163d8f9376d580288539c760968d9521db0b275dfbf8b67eb029ae70713ad3"} Dec 10 19:18:17 crc kubenswrapper[4894]: I1210 19:18:17.744210 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:18:17 crc kubenswrapper[4894]: I1210 19:18:17.744565 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:18:17 crc kubenswrapper[4894]: I1210 19:18:17.788934 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:18:17 crc kubenswrapper[4894]: I1210 19:18:17.801343 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:18:18 crc kubenswrapper[4894]: I1210 19:18:18.459651 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:18:18 crc kubenswrapper[4894]: I1210 19:18:18.459959 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:18:19 crc kubenswrapper[4894]: I1210 19:18:19.577596 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:18:19 crc kubenswrapper[4894]: I1210 19:18:19.577961 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:18:20 crc kubenswrapper[4894]: I1210 19:18:20.780814 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:18:20 crc kubenswrapper[4894]: I1210 19:18:20.780986 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:18:20 crc kubenswrapper[4894]: I1210 19:18:20.903307 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:18:25 crc kubenswrapper[4894]: I1210 19:18:25.693431 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Dec 10 19:18:30 crc kubenswrapper[4894]: I1210 19:18:30.608347 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k84gv" event={"ID":"53a93527-c2bb-4968-862c-02c60d6fe93c","Type":"ContainerStarted","Data":"c438f8d74dcb947bced6be6cd664e5c1aa2a1043d5498dc2dcf32d572c55c989"} Dec 10 19:18:30 crc kubenswrapper[4894]: I1210 19:18:30.627365 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-k84gv" podStartSLOduration=1.915796965 podStartE2EDuration="14.627340854s" podCreationTimestamp="2025-12-10 19:18:16 +0000 UTC" firstStartedPulling="2025-12-10 19:18:17.032451498 +0000 UTC m=+1379.827299300" lastFinishedPulling="2025-12-10 19:18:29.743995397 +0000 UTC m=+1392.538843189" observedRunningTime="2025-12-10 19:18:30.621884616 +0000 UTC m=+1393.416732428" watchObservedRunningTime="2025-12-10 19:18:30.627340854 +0000 UTC m=+1393.422188656" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.628954 4894 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podca7b0207-4a62-4299-aa2a-14244c283866"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podca7b0207-4a62-4299-aa2a-14244c283866] : Timed out while waiting for systemd to remove kubepods-besteffort-podca7b0207_4a62_4299_aa2a_14244c283866.slice" Dec 10 19:18:37 crc kubenswrapper[4894]: E1210 19:18:37.629501 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podca7b0207-4a62-4299-aa2a-14244c283866] : unable to destroy cgroup paths for cgroup [kubepods besteffort podca7b0207-4a62-4299-aa2a-14244c283866] : Timed out while waiting for systemd to remove kubepods-besteffort-podca7b0207_4a62_4299_aa2a_14244c283866.slice" pod="openstack/glance-default-internal-api-0" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.632314 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.694471 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.769000 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.783729 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.793553 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.795198 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.798533 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.799334 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.813411 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886119 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4rgt\" (UniqueName: \"kubernetes.io/projected/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-kube-api-access-m4rgt\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886182 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886218 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886245 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886275 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-logs\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886359 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.886379 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.989951 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990000 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4rgt\" (UniqueName: \"kubernetes.io/projected/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-kube-api-access-m4rgt\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990094 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990150 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990178 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-logs\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.990222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.996499 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-logs\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:37 crc kubenswrapper[4894]: I1210 19:18:37.996923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.007869 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.012885 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.022156 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.039116 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.040157 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.040188 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6b2877179eed34ff266ca183c106ee53ca8775eae08a086cf0505ade7a2914a4/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.067911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4rgt\" (UniqueName: \"kubernetes.io/projected/46843b9e-90e7-4d1d-9ae6-3be077fadfe5-kube-api-access-m4rgt\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.098336 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-102f2a85-63ab-4ac5-99ed-6047367290b4\") pod \"glance-default-internal-api-0\" (UID: \"46843b9e-90e7-4d1d-9ae6-3be077fadfe5\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.115029 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.662057 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:18:38 crc kubenswrapper[4894]: W1210 19:18:38.666197 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46843b9e_90e7_4d1d_9ae6_3be077fadfe5.slice/crio-ffe88acb3e455429fb7a48d6ba3cdcccff17a400ba6946dfd263904dca1bb1a8 WatchSource:0}: Error finding container ffe88acb3e455429fb7a48d6ba3cdcccff17a400ba6946dfd263904dca1bb1a8: Status 404 returned error can't find the container with id ffe88acb3e455429fb7a48d6ba3cdcccff17a400ba6946dfd263904dca1bb1a8 Dec 10 19:18:38 crc kubenswrapper[4894]: I1210 19:18:38.705149 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46843b9e-90e7-4d1d-9ae6-3be077fadfe5","Type":"ContainerStarted","Data":"ffe88acb3e455429fb7a48d6ba3cdcccff17a400ba6946dfd263904dca1bb1a8"} Dec 10 19:18:39 crc kubenswrapper[4894]: I1210 19:18:39.557804 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca7b0207-4a62-4299-aa2a-14244c283866" path="/var/lib/kubelet/pods/ca7b0207-4a62-4299-aa2a-14244c283866/volumes" Dec 10 19:18:39 crc kubenswrapper[4894]: I1210 19:18:39.720265 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46843b9e-90e7-4d1d-9ae6-3be077fadfe5","Type":"ContainerStarted","Data":"1c540040c96bf5fed1bd3002cce07349d4f8dc15b8d2fb821853a4248c5cfe4e"} Dec 10 19:18:40 crc kubenswrapper[4894]: I1210 19:18:40.732432 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46843b9e-90e7-4d1d-9ae6-3be077fadfe5","Type":"ContainerStarted","Data":"65189f2e685a965cebd09912ead620f94107800725a83122cfd8d0e618e45eda"} Dec 10 19:18:40 crc kubenswrapper[4894]: I1210 19:18:40.734070 4894 generic.go:334] "Generic (PLEG): container finished" podID="53a93527-c2bb-4968-862c-02c60d6fe93c" containerID="c438f8d74dcb947bced6be6cd664e5c1aa2a1043d5498dc2dcf32d572c55c989" exitCode=0 Dec 10 19:18:40 crc kubenswrapper[4894]: I1210 19:18:40.734109 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k84gv" event={"ID":"53a93527-c2bb-4968-862c-02c60d6fe93c","Type":"ContainerDied","Data":"c438f8d74dcb947bced6be6cd664e5c1aa2a1043d5498dc2dcf32d572c55c989"} Dec 10 19:18:40 crc kubenswrapper[4894]: I1210 19:18:40.758163 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.758139768 podStartE2EDuration="3.758139768s" podCreationTimestamp="2025-12-10 19:18:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:40.75269527 +0000 UTC m=+1403.547543082" watchObservedRunningTime="2025-12-10 19:18:40.758139768 +0000 UTC m=+1403.552987570" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.219127 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.399613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-scripts\") pod \"53a93527-c2bb-4968-862c-02c60d6fe93c\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.399795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-combined-ca-bundle\") pod \"53a93527-c2bb-4968-862c-02c60d6fe93c\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.400149 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-config-data\") pod \"53a93527-c2bb-4968-862c-02c60d6fe93c\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.400807 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjc4v\" (UniqueName: \"kubernetes.io/projected/53a93527-c2bb-4968-862c-02c60d6fe93c-kube-api-access-wjc4v\") pod \"53a93527-c2bb-4968-862c-02c60d6fe93c\" (UID: \"53a93527-c2bb-4968-862c-02c60d6fe93c\") " Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.406189 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-scripts" (OuterVolumeSpecName: "scripts") pod "53a93527-c2bb-4968-862c-02c60d6fe93c" (UID: "53a93527-c2bb-4968-862c-02c60d6fe93c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.406216 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a93527-c2bb-4968-862c-02c60d6fe93c-kube-api-access-wjc4v" (OuterVolumeSpecName: "kube-api-access-wjc4v") pod "53a93527-c2bb-4968-862c-02c60d6fe93c" (UID: "53a93527-c2bb-4968-862c-02c60d6fe93c"). InnerVolumeSpecName "kube-api-access-wjc4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.442772 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53a93527-c2bb-4968-862c-02c60d6fe93c" (UID: "53a93527-c2bb-4968-862c-02c60d6fe93c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.454051 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-config-data" (OuterVolumeSpecName: "config-data") pod "53a93527-c2bb-4968-862c-02c60d6fe93c" (UID: "53a93527-c2bb-4968-862c-02c60d6fe93c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.505508 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.505559 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjc4v\" (UniqueName: \"kubernetes.io/projected/53a93527-c2bb-4968-862c-02c60d6fe93c-kube-api-access-wjc4v\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.505581 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.505598 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53a93527-c2bb-4968-862c-02c60d6fe93c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.763547 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k84gv" event={"ID":"53a93527-c2bb-4968-862c-02c60d6fe93c","Type":"ContainerDied","Data":"3b163d8f9376d580288539c760968d9521db0b275dfbf8b67eb029ae70713ad3"} Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.763592 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b163d8f9376d580288539c760968d9521db0b275dfbf8b67eb029ae70713ad3" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.763643 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k84gv" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.953107 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:18:42 crc kubenswrapper[4894]: E1210 19:18:42.953557 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a93527-c2bb-4968-862c-02c60d6fe93c" containerName="nova-cell0-conductor-db-sync" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.953580 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a93527-c2bb-4968-862c-02c60d6fe93c" containerName="nova-cell0-conductor-db-sync" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.953762 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a93527-c2bb-4968-862c-02c60d6fe93c" containerName="nova-cell0-conductor-db-sync" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.954521 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.956422 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2479m" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.957392 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 19:18:42 crc kubenswrapper[4894]: I1210 19:18:42.968318 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.121611 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vkk9\" (UniqueName: \"kubernetes.io/projected/e75c654f-e152-4544-b6bd-a93466dcb0fc-kube-api-access-9vkk9\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.121690 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75c654f-e152-4544-b6bd-a93466dcb0fc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.121739 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75c654f-e152-4544-b6bd-a93466dcb0fc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.224735 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vkk9\" (UniqueName: \"kubernetes.io/projected/e75c654f-e152-4544-b6bd-a93466dcb0fc-kube-api-access-9vkk9\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.224835 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75c654f-e152-4544-b6bd-a93466dcb0fc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.224920 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75c654f-e152-4544-b6bd-a93466dcb0fc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.231330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e75c654f-e152-4544-b6bd-a93466dcb0fc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.231694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e75c654f-e152-4544-b6bd-a93466dcb0fc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.248800 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vkk9\" (UniqueName: \"kubernetes.io/projected/e75c654f-e152-4544-b6bd-a93466dcb0fc-kube-api-access-9vkk9\") pod \"nova-cell0-conductor-0\" (UID: \"e75c654f-e152-4544-b6bd-a93466dcb0fc\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.270448 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.766017 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.779326 4894 generic.go:334] "Generic (PLEG): container finished" podID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerID="4062b40e399e0ec3e7808e5263fec0bf8ac81cf9762ec345e153da7a5140f5d0" exitCode=137 Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.779402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerDied","Data":"4062b40e399e0ec3e7808e5263fec0bf8ac81cf9762ec345e153da7a5140f5d0"} Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.780572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e75c654f-e152-4544-b6bd-a93466dcb0fc","Type":"ContainerStarted","Data":"8381f4a6e576e851e0b90cf97ec559a3860d65382a2fb1125a22a67b90291ca6"} Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.818082 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.950541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-sg-core-conf-yaml\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.950984 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7x4m\" (UniqueName: \"kubernetes.io/projected/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-kube-api-access-b7x4m\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.951048 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-scripts\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.951118 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-run-httpd\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.951154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-config-data\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.951194 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-log-httpd\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.951317 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-combined-ca-bundle\") pod \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\" (UID: \"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45\") " Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.959168 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.959616 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.959752 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-scripts" (OuterVolumeSpecName: "scripts") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.959978 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-kube-api-access-b7x4m" (OuterVolumeSpecName: "kube-api-access-b7x4m") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "kube-api-access-b7x4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.961087 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7x4m\" (UniqueName: \"kubernetes.io/projected/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-kube-api-access-b7x4m\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.961249 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.961282 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:43 crc kubenswrapper[4894]: I1210 19:18:43.961295 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.010059 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.045075 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.063581 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.063627 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.077209 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-config-data" (OuterVolumeSpecName: "config-data") pod "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" (UID: "0c5f2bfd-a3f9-4952-b99b-c896cd58cd45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.166043 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.798334 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e75c654f-e152-4544-b6bd-a93466dcb0fc","Type":"ContainerStarted","Data":"9af606b26cdb572f4def6c3a3ab7adce915981873362e334d2264ff56a06722d"} Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.798461 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.807482 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c5f2bfd-a3f9-4952-b99b-c896cd58cd45","Type":"ContainerDied","Data":"732b6fa634b14916afc415876f659e3ae71bf85a731b0d0c10b71e7856b55b3b"} Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.807564 4894 scope.go:117] "RemoveContainer" containerID="4062b40e399e0ec3e7808e5263fec0bf8ac81cf9762ec345e153da7a5140f5d0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.807622 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.858265 4894 scope.go:117] "RemoveContainer" containerID="6e7e9447eb5a49a984363a368a3ffdf30e7216d7a0c2b036d48b7df7094d5c4b" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.869628 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.869600786 podStartE2EDuration="2.869600786s" podCreationTimestamp="2025-12-10 19:18:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:44.832508271 +0000 UTC m=+1407.627356063" watchObservedRunningTime="2025-12-10 19:18:44.869600786 +0000 UTC m=+1407.664448608" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.886567 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.904389 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.916347 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:44 crc kubenswrapper[4894]: E1210 19:18:44.917143 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="proxy-httpd" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917180 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="proxy-httpd" Dec 10 19:18:44 crc kubenswrapper[4894]: E1210 19:18:44.917209 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-notification-agent" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917222 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-notification-agent" Dec 10 19:18:44 crc kubenswrapper[4894]: E1210 19:18:44.917261 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-central-agent" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917272 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-central-agent" Dec 10 19:18:44 crc kubenswrapper[4894]: E1210 19:18:44.917298 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="sg-core" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917308 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="sg-core" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917602 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-notification-agent" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917638 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="ceilometer-central-agent" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917654 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="sg-core" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.917689 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" containerName="proxy-httpd" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.918047 4894 scope.go:117] "RemoveContainer" containerID="81c920761368c090c0dafbc881b940ffc060c03579a7436c924bd2bb9cee1ee7" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.921241 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.924174 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.924484 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.928900 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.980536 4894 scope.go:117] "RemoveContainer" containerID="f620f337eb625e5206af5aaf50840b89eb91103c654cfeeab5dfd732b56c32d8" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984249 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984326 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-run-httpd\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984357 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-949s8\" (UniqueName: \"kubernetes.io/projected/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-kube-api-access-949s8\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984425 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-scripts\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-log-httpd\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:44 crc kubenswrapper[4894]: I1210 19:18:44.984570 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-config-data\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-log-httpd\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086273 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-config-data\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086354 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086395 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-run-httpd\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086418 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086441 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-949s8\" (UniqueName: \"kubernetes.io/projected/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-kube-api-access-949s8\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.086468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-scripts\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.089815 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-log-httpd\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.090294 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.090524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-run-httpd\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.090892 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-scripts\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.092459 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.093109 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-config-data\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.110889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-949s8\" (UniqueName: \"kubernetes.io/projected/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-kube-api-access-949s8\") pod \"ceilometer-0\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.268006 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.558195 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5f2bfd-a3f9-4952-b99b-c896cd58cd45" path="/var/lib/kubelet/pods/0c5f2bfd-a3f9-4952-b99b-c896cd58cd45/volumes" Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.740586 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:18:45 crc kubenswrapper[4894]: I1210 19:18:45.819633 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerStarted","Data":"f4d569d4ad1937809089642c84bc548ae90ec23332cc57bda552d7ca80df370c"} Dec 10 19:18:46 crc kubenswrapper[4894]: I1210 19:18:46.832759 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerStarted","Data":"751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d"} Dec 10 19:18:47 crc kubenswrapper[4894]: I1210 19:18:47.847695 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerStarted","Data":"fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5"} Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.116019 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.116077 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.156421 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.196573 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.302144 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.750598 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-pglr9"] Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.756021 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.759922 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.760172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.782974 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pglr9"] Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.865581 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m94rh\" (UniqueName: \"kubernetes.io/projected/a57e8fe0-db28-42be-aa0d-f39f751d7897-kube-api-access-m94rh\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.865797 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-scripts\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.865826 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-config-data\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.865937 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.882940 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerStarted","Data":"d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e"} Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.883001 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.883223 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.967605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-scripts\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.967811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-config-data\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.967871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.967946 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m94rh\" (UniqueName: \"kubernetes.io/projected/a57e8fe0-db28-42be-aa0d-f39f751d7897-kube-api-access-m94rh\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.967759 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.970809 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.978188 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.978265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-config-data\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.984377 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-scripts\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:48 crc kubenswrapper[4894]: I1210 19:18:48.988158 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.005369 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.017796 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.019811 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.031191 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m94rh\" (UniqueName: \"kubernetes.io/projected/a57e8fe0-db28-42be-aa0d-f39f751d7897-kube-api-access-m94rh\") pod \"nova-cell0-cell-mapping-pglr9\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.031197 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.040378 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.069837 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.069910 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.070013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56kz8\" (UniqueName: \"kubernetes.io/projected/02f732ef-d8c7-437b-8079-e6da9a0537e9-kube-api-access-56kz8\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.078396 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.148921 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.150435 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.163044 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.171837 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56kz8\" (UniqueName: \"kubernetes.io/projected/02f732ef-d8c7-437b-8079-e6da9a0537e9-kube-api-access-56kz8\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.171929 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.171953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4qmn\" (UniqueName: \"kubernetes.io/projected/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-kube-api-access-h4qmn\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.172010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.172027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-config-data\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.172060 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-logs\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.172089 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.187601 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.215972 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c9cb78d75-nqzt2"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.217650 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.234606 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.239886 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c9cb78d75-nqzt2"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.249368 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56kz8\" (UniqueName: \"kubernetes.io/projected/02f732ef-d8c7-437b-8079-e6da9a0537e9-kube-api-access-56kz8\") pod \"nova-cell1-novncproxy-0\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.273859 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.273939 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-nb\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.273970 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-sb\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.273993 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k89kz\" (UniqueName: \"kubernetes.io/projected/ca6401e5-caa0-49b6-a281-b74b74190143-kube-api-access-k89kz\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274011 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-svc\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274033 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-swift-storage-0\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274052 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-config-data\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274099 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4qmn\" (UniqueName: \"kubernetes.io/projected/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-kube-api-access-h4qmn\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274169 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65qsx\" (UniqueName: \"kubernetes.io/projected/be401a3a-dd33-46cc-9e88-f402415a996d-kube-api-access-65qsx\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274196 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-config-data\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274219 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-logs\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.274245 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-config\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.280539 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.283299 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-logs\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.286441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-config-data\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.300827 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.306900 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.317544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4qmn\" (UniqueName: \"kubernetes.io/projected/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-kube-api-access-h4qmn\") pod \"nova-metadata-0\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.320290 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.321826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.323668 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.334302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.348900 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.375802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k89kz\" (UniqueName: \"kubernetes.io/projected/ca6401e5-caa0-49b6-a281-b74b74190143-kube-api-access-k89kz\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.375858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-svc\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.375884 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-swift-storage-0\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.375903 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-config-data\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.375962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65qsx\" (UniqueName: \"kubernetes.io/projected/be401a3a-dd33-46cc-9e88-f402415a996d-kube-api-access-65qsx\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.376006 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsdd6\" (UniqueName: \"kubernetes.io/projected/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-kube-api-access-gsdd6\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.376063 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-config\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.419692 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-svc\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.420105 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-config\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.422962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-config-data\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.423331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.423541 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.423727 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-logs\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.423935 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-nb\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.424263 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-sb\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.438825 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-sb\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.441981 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-config-data\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.445830 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.447250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k89kz\" (UniqueName: \"kubernetes.io/projected/ca6401e5-caa0-49b6-a281-b74b74190143-kube-api-access-k89kz\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.447647 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-swift-storage-0\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.451921 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-nb\") pod \"dnsmasq-dns-7c9cb78d75-nqzt2\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.453315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65qsx\" (UniqueName: \"kubernetes.io/projected/be401a3a-dd33-46cc-9e88-f402415a996d-kube-api-access-65qsx\") pod \"nova-scheduler-0\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.542224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.542536 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-logs\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.542758 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsdd6\" (UniqueName: \"kubernetes.io/projected/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-kube-api-access-gsdd6\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.542823 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-config-data\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.549406 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-logs\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.549855 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.551530 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-config-data\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.570106 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsdd6\" (UniqueName: \"kubernetes.io/projected/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-kube-api-access-gsdd6\") pod \"nova-api-0\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.577201 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.577256 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.577300 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.578036 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c62314c4a97570eb40f37c7f81fec8ac3dc6ad9954e24945138dbb31ee1ce97"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.578083 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://5c62314c4a97570eb40f37c7f81fec8ac3dc6ad9954e24945138dbb31ee1ce97" gracePeriod=600 Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.729172 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.740358 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.760489 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.837759 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-pglr9"] Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.930695 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="5c62314c4a97570eb40f37c7f81fec8ac3dc6ad9954e24945138dbb31ee1ce97" exitCode=0 Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.932471 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"5c62314c4a97570eb40f37c7f81fec8ac3dc6ad9954e24945138dbb31ee1ce97"} Dec 10 19:18:49 crc kubenswrapper[4894]: I1210 19:18:49.932554 4894 scope.go:117] "RemoveContainer" containerID="74b5960f63532f14bffacaae454d958ec29335c9f8adf904077f837019704bae" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.032123 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8ntxv"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.034832 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.039906 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.042296 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.080769 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.097207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8ntxv"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.162191 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.172982 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.173037 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-scripts\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.173226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2czt\" (UniqueName: \"kubernetes.io/projected/896eec35-0551-4bf5-9afa-179674319fdd-kube-api-access-c2czt\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.173632 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-config-data\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: W1210 19:18:50.187982 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d11f74a_c4c6_48f7_9c36_29777f70f1a2.slice/crio-7ee6884cc1c6b051eff4d2bdcdd013ba9eac7c4915ee5ba60e02e28f59ff0acd WatchSource:0}: Error finding container 7ee6884cc1c6b051eff4d2bdcdd013ba9eac7c4915ee5ba60e02e28f59ff0acd: Status 404 returned error can't find the container with id 7ee6884cc1c6b051eff4d2bdcdd013ba9eac7c4915ee5ba60e02e28f59ff0acd Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.247008 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2wmmj"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.249153 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.274984 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2czt\" (UniqueName: \"kubernetes.io/projected/896eec35-0551-4bf5-9afa-179674319fdd-kube-api-access-c2czt\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.275141 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-config-data\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.275177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.275201 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-scripts\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.276937 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2wmmj"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.280971 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-scripts\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.286413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.295357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2czt\" (UniqueName: \"kubernetes.io/projected/896eec35-0551-4bf5-9afa-179674319fdd-kube-api-access-c2czt\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.314366 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-config-data\") pod \"nova-cell1-conductor-db-sync-8ntxv\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.378144 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-utilities\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.380001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cnmv\" (UniqueName: \"kubernetes.io/projected/74a13406-618a-4f69-b9de-fafec8c37346-kube-api-access-7cnmv\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.380113 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-catalog-content\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.420201 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.483411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-utilities\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.483477 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cnmv\" (UniqueName: \"kubernetes.io/projected/74a13406-618a-4f69-b9de-fafec8c37346-kube-api-access-7cnmv\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.483517 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-catalog-content\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.483933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-catalog-content\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.484157 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-utilities\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.599710 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.626643 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cnmv\" (UniqueName: \"kubernetes.io/projected/74a13406-618a-4f69-b9de-fafec8c37346-kube-api-access-7cnmv\") pod \"redhat-operators-2wmmj\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.647519 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c9cb78d75-nqzt2"] Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.651346 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:18:50 crc kubenswrapper[4894]: I1210 19:18:50.901309 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.061128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d11f74a-c4c6-48f7-9c36-29777f70f1a2","Type":"ContainerStarted","Data":"7ee6884cc1c6b051eff4d2bdcdd013ba9eac7c4915ee5ba60e02e28f59ff0acd"} Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.073955 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c","Type":"ContainerStarted","Data":"c6487005a7ffdc2a9473d4054dc2e6aebb73bb5964098812d4b4917db7807af6"} Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.085154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4"} Dec 10 19:18:51 crc kubenswrapper[4894]: W1210 19:18:51.122307 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe401a3a_dd33_46cc_9e88_f402415a996d.slice/crio-a09fb5383ecb24e1b6483575e7005eb12b53c7d8bfa02d7b57111600efc773b2 WatchSource:0}: Error finding container a09fb5383ecb24e1b6483575e7005eb12b53c7d8bfa02d7b57111600efc773b2: Status 404 returned error can't find the container with id a09fb5383ecb24e1b6483575e7005eb12b53c7d8bfa02d7b57111600efc773b2 Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.163093 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" event={"ID":"ca6401e5-caa0-49b6-a281-b74b74190143","Type":"ContainerStarted","Data":"62e84fc7c26de3760876f5c6883205ac0e76016ea76ae6f63a44f1c3097a51ef"} Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.270425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"02f732ef-d8c7-437b-8079-e6da9a0537e9","Type":"ContainerStarted","Data":"6f5e9c01e8daead9f91b78c43360b20f3bf3ced80d0add9c4d6123b6d020e35c"} Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.308771 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerStarted","Data":"d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7"} Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.310450 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.350599 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.350824 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.352325 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pglr9" event={"ID":"a57e8fe0-db28-42be-aa0d-f39f751d7897","Type":"ContainerStarted","Data":"77d9cbdb0a2508e53f6078e52294459cd5cd31cf40ae0b984d0f3b61bf845628"} Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.389007 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.340142366 podStartE2EDuration="7.371198728s" podCreationTimestamp="2025-12-10 19:18:44 +0000 UTC" firstStartedPulling="2025-12-10 19:18:45.738579464 +0000 UTC m=+1408.533427256" lastFinishedPulling="2025-12-10 19:18:49.769635826 +0000 UTC m=+1412.564483618" observedRunningTime="2025-12-10 19:18:51.344808024 +0000 UTC m=+1414.139655836" watchObservedRunningTime="2025-12-10 19:18:51.371198728 +0000 UTC m=+1414.166046520" Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.404433 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-pglr9" podStartSLOduration=3.404413837 podStartE2EDuration="3.404413837s" podCreationTimestamp="2025-12-10 19:18:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:51.369661566 +0000 UTC m=+1414.164509378" watchObservedRunningTime="2025-12-10 19:18:51.404413837 +0000 UTC m=+1414.199261619" Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.598579 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8ntxv"] Dec 10 19:18:51 crc kubenswrapper[4894]: I1210 19:18:51.650245 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2wmmj"] Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.389515 4894 generic.go:334] "Generic (PLEG): container finished" podID="74a13406-618a-4f69-b9de-fafec8c37346" containerID="0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3" exitCode=0 Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.390161 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2wmmj" event={"ID":"74a13406-618a-4f69-b9de-fafec8c37346","Type":"ContainerDied","Data":"0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.390196 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2wmmj" event={"ID":"74a13406-618a-4f69-b9de-fafec8c37346","Type":"ContainerStarted","Data":"aaf19152b19f8165c3194dbb89064ed8def1bec63ea36ea5b7ce5032bcf100fc"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.393831 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pglr9" event={"ID":"a57e8fe0-db28-42be-aa0d-f39f751d7897","Type":"ContainerStarted","Data":"02f90fc7a4bd62f67dfd1457046169268063d085c23f8c34c4abcbe11a807ed8"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.404213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" event={"ID":"896eec35-0551-4bf5-9afa-179674319fdd","Type":"ContainerStarted","Data":"f60fb0e5527c8ea654ab78a3f1641d06d5e2890ada4ec13c23585124cea12666"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.404393 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" event={"ID":"896eec35-0551-4bf5-9afa-179674319fdd","Type":"ContainerStarted","Data":"8de1ccdce0ee7ce42b9bfa842f9d7476a81567edb99d09663e6e3cf58b5fa1dc"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.413976 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be401a3a-dd33-46cc-9e88-f402415a996d","Type":"ContainerStarted","Data":"a09fb5383ecb24e1b6483575e7005eb12b53c7d8bfa02d7b57111600efc773b2"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.422363 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" podStartSLOduration=3.422347628 podStartE2EDuration="3.422347628s" podCreationTimestamp="2025-12-10 19:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:52.417086106 +0000 UTC m=+1415.211933898" watchObservedRunningTime="2025-12-10 19:18:52.422347628 +0000 UTC m=+1415.217195420" Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.423147 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca6401e5-caa0-49b6-a281-b74b74190143" containerID="9710d519a07e8e8982454ed02387beb1dde89fcba3bab3ab6c485542a59469ea" exitCode=0 Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.423914 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" event={"ID":"ca6401e5-caa0-49b6-a281-b74b74190143","Type":"ContainerDied","Data":"9710d519a07e8e8982454ed02387beb1dde89fcba3bab3ab6c485542a59469ea"} Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.713253 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.713597 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.781987 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.830049 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:18:52 crc kubenswrapper[4894]: I1210 19:18:52.856973 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:18:53 crc kubenswrapper[4894]: I1210 19:18:53.438179 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" event={"ID":"ca6401e5-caa0-49b6-a281-b74b74190143","Type":"ContainerStarted","Data":"cc469c503b32241dd808394a83a6e9ef61915a403dc1a2be8f9537d87b7327d7"} Dec 10 19:18:53 crc kubenswrapper[4894]: I1210 19:18:53.469373 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" podStartSLOduration=4.469354956 podStartE2EDuration="4.469354956s" podCreationTimestamp="2025-12-10 19:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.460438185 +0000 UTC m=+1416.255285977" watchObservedRunningTime="2025-12-10 19:18:53.469354956 +0000 UTC m=+1416.264202748" Dec 10 19:18:54 crc kubenswrapper[4894]: I1210 19:18:54.450093 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.472548 4894 generic.go:334] "Generic (PLEG): container finished" podID="74a13406-618a-4f69-b9de-fafec8c37346" containerID="b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd" exitCode=0 Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.472621 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2wmmj" event={"ID":"74a13406-618a-4f69-b9de-fafec8c37346","Type":"ContainerDied","Data":"b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd"} Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.478481 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d11f74a-c4c6-48f7-9c36-29777f70f1a2","Type":"ContainerStarted","Data":"7dfc8571dd2e3aa6cc2a3214ab9544bf6d74793321d74b4c06bb9cbf41a9fc89"} Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.481941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c","Type":"ContainerStarted","Data":"effa0119c3772f8bf8b70617cd7d9b4926738127b9f700c404bbba1012e14581"} Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.486655 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be401a3a-dd33-46cc-9e88-f402415a996d","Type":"ContainerStarted","Data":"39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27"} Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.489938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"02f732ef-d8c7-437b-8079-e6da9a0537e9","Type":"ContainerStarted","Data":"378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2"} Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.490066 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="02f732ef-d8c7-437b-8079-e6da9a0537e9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2" gracePeriod=30 Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.551949 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.576163318 podStartE2EDuration="8.551928198s" podCreationTimestamp="2025-12-10 19:18:48 +0000 UTC" firstStartedPulling="2025-12-10 19:18:50.141118284 +0000 UTC m=+1412.935966076" lastFinishedPulling="2025-12-10 19:18:55.116883154 +0000 UTC m=+1417.911730956" observedRunningTime="2025-12-10 19:18:56.527773653 +0000 UTC m=+1419.322621455" watchObservedRunningTime="2025-12-10 19:18:56.551928198 +0000 UTC m=+1419.346775990" Dec 10 19:18:56 crc kubenswrapper[4894]: I1210 19:18:56.554127 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.643332292 podStartE2EDuration="7.554115197s" podCreationTimestamp="2025-12-10 19:18:49 +0000 UTC" firstStartedPulling="2025-12-10 19:18:51.210995121 +0000 UTC m=+1414.005842913" lastFinishedPulling="2025-12-10 19:18:55.121778026 +0000 UTC m=+1417.916625818" observedRunningTime="2025-12-10 19:18:56.545580146 +0000 UTC m=+1419.340427948" watchObservedRunningTime="2025-12-10 19:18:56.554115197 +0000 UTC m=+1419.348962989" Dec 10 19:18:57 crc kubenswrapper[4894]: I1210 19:18:57.509376 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c","Type":"ContainerStarted","Data":"4953d82118e9f674f5a4f0027f35b08545f0ff14e5451273d1d2db5d670a2bd6"} Dec 10 19:18:57 crc kubenswrapper[4894]: I1210 19:18:57.514986 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-log" containerID="cri-o://7dfc8571dd2e3aa6cc2a3214ab9544bf6d74793321d74b4c06bb9cbf41a9fc89" gracePeriod=30 Dec 10 19:18:57 crc kubenswrapper[4894]: I1210 19:18:57.515172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d11f74a-c4c6-48f7-9c36-29777f70f1a2","Type":"ContainerStarted","Data":"11270b111456200dd77ce68766d8fbbe654cdf04a8c7e8718c9d8c79a996f1de"} Dec 10 19:18:57 crc kubenswrapper[4894]: I1210 19:18:57.515227 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-metadata" containerID="cri-o://11270b111456200dd77ce68766d8fbbe654cdf04a8c7e8718c9d8c79a996f1de" gracePeriod=30 Dec 10 19:18:57 crc kubenswrapper[4894]: I1210 19:18:57.536947 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.83162424 podStartE2EDuration="8.536924687s" podCreationTimestamp="2025-12-10 19:18:49 +0000 UTC" firstStartedPulling="2025-12-10 19:18:50.413383045 +0000 UTC m=+1413.208230837" lastFinishedPulling="2025-12-10 19:18:55.118683492 +0000 UTC m=+1417.913531284" observedRunningTime="2025-12-10 19:18:57.524200772 +0000 UTC m=+1420.319048584" watchObservedRunningTime="2025-12-10 19:18:57.536924687 +0000 UTC m=+1420.331772489" Dec 10 19:18:57 crc kubenswrapper[4894]: I1210 19:18:57.563157 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.656293003 podStartE2EDuration="9.563139257s" podCreationTimestamp="2025-12-10 19:18:48 +0000 UTC" firstStartedPulling="2025-12-10 19:18:50.214033548 +0000 UTC m=+1413.008881340" lastFinishedPulling="2025-12-10 19:18:55.120879802 +0000 UTC m=+1417.915727594" observedRunningTime="2025-12-10 19:18:57.556170208 +0000 UTC m=+1420.351018010" watchObservedRunningTime="2025-12-10 19:18:57.563139257 +0000 UTC m=+1420.357987049" Dec 10 19:18:58 crc kubenswrapper[4894]: I1210 19:18:58.524690 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2wmmj" event={"ID":"74a13406-618a-4f69-b9de-fafec8c37346","Type":"ContainerStarted","Data":"7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8"} Dec 10 19:18:58 crc kubenswrapper[4894]: I1210 19:18:58.529563 4894 generic.go:334] "Generic (PLEG): container finished" podID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerID="11270b111456200dd77ce68766d8fbbe654cdf04a8c7e8718c9d8c79a996f1de" exitCode=0 Dec 10 19:18:58 crc kubenswrapper[4894]: I1210 19:18:58.529595 4894 generic.go:334] "Generic (PLEG): container finished" podID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerID="7dfc8571dd2e3aa6cc2a3214ab9544bf6d74793321d74b4c06bb9cbf41a9fc89" exitCode=143 Dec 10 19:18:58 crc kubenswrapper[4894]: I1210 19:18:58.530329 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d11f74a-c4c6-48f7-9c36-29777f70f1a2","Type":"ContainerDied","Data":"11270b111456200dd77ce68766d8fbbe654cdf04a8c7e8718c9d8c79a996f1de"} Dec 10 19:18:58 crc kubenswrapper[4894]: I1210 19:18:58.530364 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d11f74a-c4c6-48f7-9c36-29777f70f1a2","Type":"ContainerDied","Data":"7dfc8571dd2e3aa6cc2a3214ab9544bf6d74793321d74b4c06bb9cbf41a9fc89"} Dec 10 19:18:58 crc kubenswrapper[4894]: I1210 19:18:58.562940 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2wmmj" podStartSLOduration=5.128529369 podStartE2EDuration="8.562902816s" podCreationTimestamp="2025-12-10 19:18:50 +0000 UTC" firstStartedPulling="2025-12-10 19:18:53.846269291 +0000 UTC m=+1416.641117083" lastFinishedPulling="2025-12-10 19:18:57.280642738 +0000 UTC m=+1420.075490530" observedRunningTime="2025-12-10 19:18:58.54718902 +0000 UTC m=+1421.342036812" watchObservedRunningTime="2025-12-10 19:18:58.562902816 +0000 UTC m=+1421.357750678" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.308314 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.335896 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.335959 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.730417 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.730676 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.743025 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.749327 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.761415 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.761489 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.776622 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.838033 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86d9875b97-pnsfc"] Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.838262 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerName="dnsmasq-dns" containerID="cri-o://fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd" gracePeriod=10 Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.899399 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4qmn\" (UniqueName: \"kubernetes.io/projected/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-kube-api-access-h4qmn\") pod \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.899666 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-config-data\") pod \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.899759 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-logs\") pod \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.900052 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-combined-ca-bundle\") pod \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\" (UID: \"5d11f74a-c4c6-48f7-9c36-29777f70f1a2\") " Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.900401 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-logs" (OuterVolumeSpecName: "logs") pod "5d11f74a-c4c6-48f7-9c36-29777f70f1a2" (UID: "5d11f74a-c4c6-48f7-9c36-29777f70f1a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.901100 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.905992 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-kube-api-access-h4qmn" (OuterVolumeSpecName: "kube-api-access-h4qmn") pod "5d11f74a-c4c6-48f7-9c36-29777f70f1a2" (UID: "5d11f74a-c4c6-48f7-9c36-29777f70f1a2"). InnerVolumeSpecName "kube-api-access-h4qmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.938979 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-config-data" (OuterVolumeSpecName: "config-data") pod "5d11f74a-c4c6-48f7-9c36-29777f70f1a2" (UID: "5d11f74a-c4c6-48f7-9c36-29777f70f1a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:59 crc kubenswrapper[4894]: I1210 19:18:59.941007 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d11f74a-c4c6-48f7-9c36-29777f70f1a2" (UID: "5d11f74a-c4c6-48f7-9c36-29777f70f1a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.010784 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4qmn\" (UniqueName: \"kubernetes.io/projected/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-kube-api-access-h4qmn\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.011315 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.011327 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d11f74a-c4c6-48f7-9c36-29777f70f1a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.500746 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.556062 4894 generic.go:334] "Generic (PLEG): container finished" podID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerID="fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd" exitCode=0 Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.556111 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.556137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" event={"ID":"b4ef5729-6000-469a-bf66-a53c71bda7e0","Type":"ContainerDied","Data":"fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd"} Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.556161 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86d9875b97-pnsfc" event={"ID":"b4ef5729-6000-469a-bf66-a53c71bda7e0","Type":"ContainerDied","Data":"4fd6031d77791ddf20799fa26f085733601a0e488268d7cdb3fd19af8a98ec3e"} Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.556178 4894 scope.go:117] "RemoveContainer" containerID="fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.567490 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.567776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d11f74a-c4c6-48f7-9c36-29777f70f1a2","Type":"ContainerDied","Data":"7ee6884cc1c6b051eff4d2bdcdd013ba9eac7c4915ee5ba60e02e28f59ff0acd"} Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.625559 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-swift-storage-0\") pod \"b4ef5729-6000-469a-bf66-a53c71bda7e0\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.625643 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-svc\") pod \"b4ef5729-6000-469a-bf66-a53c71bda7e0\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.625709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-nb\") pod \"b4ef5729-6000-469a-bf66-a53c71bda7e0\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.625834 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m7zv\" (UniqueName: \"kubernetes.io/projected/b4ef5729-6000-469a-bf66-a53c71bda7e0-kube-api-access-2m7zv\") pod \"b4ef5729-6000-469a-bf66-a53c71bda7e0\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.625895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-sb\") pod \"b4ef5729-6000-469a-bf66-a53c71bda7e0\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.625933 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-config\") pod \"b4ef5729-6000-469a-bf66-a53c71bda7e0\" (UID: \"b4ef5729-6000-469a-bf66-a53c71bda7e0\") " Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.663268 4894 scope.go:117] "RemoveContainer" containerID="13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.666489 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4ef5729-6000-469a-bf66-a53c71bda7e0-kube-api-access-2m7zv" (OuterVolumeSpecName: "kube-api-access-2m7zv") pod "b4ef5729-6000-469a-bf66-a53c71bda7e0" (UID: "b4ef5729-6000-469a-bf66-a53c71bda7e0"). InnerVolumeSpecName "kube-api-access-2m7zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.666786 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.666944 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.702110 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.734942 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.736343 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m7zv\" (UniqueName: \"kubernetes.io/projected/b4ef5729-6000-469a-bf66-a53c71bda7e0-kube-api-access-2m7zv\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.745773 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b4ef5729-6000-469a-bf66-a53c71bda7e0" (UID: "b4ef5729-6000-469a-bf66-a53c71bda7e0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.765953 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-config" (OuterVolumeSpecName: "config") pod "b4ef5729-6000-469a-bf66-a53c71bda7e0" (UID: "b4ef5729-6000-469a-bf66-a53c71bda7e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.778526 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4ef5729-6000-469a-bf66-a53c71bda7e0" (UID: "b4ef5729-6000-469a-bf66-a53c71bda7e0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.783933 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.813962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4ef5729-6000-469a-bf66-a53c71bda7e0" (UID: "b4ef5729-6000-469a-bf66-a53c71bda7e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.824518 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4ef5729-6000-469a-bf66-a53c71bda7e0" (UID: "b4ef5729-6000-469a-bf66-a53c71bda7e0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.828644 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:00 crc kubenswrapper[4894]: E1210 19:19:00.829162 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-log" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829183 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-log" Dec 10 19:19:00 crc kubenswrapper[4894]: E1210 19:19:00.829222 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-metadata" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829231 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-metadata" Dec 10 19:19:00 crc kubenswrapper[4894]: E1210 19:19:00.829324 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerName="init" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829331 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerName="init" Dec 10 19:19:00 crc kubenswrapper[4894]: E1210 19:19:00.829357 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerName="dnsmasq-dns" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829363 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerName="dnsmasq-dns" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829579 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" containerName="dnsmasq-dns" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829591 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-metadata" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.829601 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" containerName="nova-metadata-log" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.830763 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.841356 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.841566 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.843472 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.843492 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.843503 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.843512 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.843519 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ef5729-6000-469a-bf66-a53c71bda7e0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.845811 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.210:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.846092 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.210:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.881482 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.926327 4894 scope.go:117] "RemoveContainer" containerID="fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd" Dec 10 19:19:00 crc kubenswrapper[4894]: E1210 19:19:00.930531 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd\": container with ID starting with fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd not found: ID does not exist" containerID="fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.930596 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd"} err="failed to get container status \"fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd\": rpc error: code = NotFound desc = could not find container \"fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd\": container with ID starting with fe3124c6438121efdcd3c0fdd2a79fc2146b7106eb340cea2e479c4b87e2b5bd not found: ID does not exist" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.930628 4894 scope.go:117] "RemoveContainer" containerID="13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0" Dec 10 19:19:00 crc kubenswrapper[4894]: E1210 19:19:00.931925 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0\": container with ID starting with 13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0 not found: ID does not exist" containerID="13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.931991 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0"} err="failed to get container status \"13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0\": rpc error: code = NotFound desc = could not find container \"13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0\": container with ID starting with 13a9bf3698190b00bc07dc5334c84c223cc53f9f8c245163833fdee3dbc77ce0 not found: ID does not exist" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.932015 4894 scope.go:117] "RemoveContainer" containerID="11270b111456200dd77ce68766d8fbbe654cdf04a8c7e8718c9d8c79a996f1de" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.971423 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-config-data\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.971672 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hsfm\" (UniqueName: \"kubernetes.io/projected/0872c419-8add-4f2c-9216-859b39aff2ec-kube-api-access-5hsfm\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.971839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.971987 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0872c419-8add-4f2c-9216-859b39aff2ec-logs\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.972196 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.973558 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86d9875b97-pnsfc"] Dec 10 19:19:00 crc kubenswrapper[4894]: I1210 19:19:00.980208 4894 scope.go:117] "RemoveContainer" containerID="7dfc8571dd2e3aa6cc2a3214ab9544bf6d74793321d74b4c06bb9cbf41a9fc89" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.009197 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86d9875b97-pnsfc"] Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.074382 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.074493 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0872c419-8add-4f2c-9216-859b39aff2ec-logs\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.074555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.074628 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-config-data\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.074701 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hsfm\" (UniqueName: \"kubernetes.io/projected/0872c419-8add-4f2c-9216-859b39aff2ec-kube-api-access-5hsfm\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.075031 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0872c419-8add-4f2c-9216-859b39aff2ec-logs\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.080043 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-config-data\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.080328 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.080902 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.093813 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hsfm\" (UniqueName: \"kubernetes.io/projected/0872c419-8add-4f2c-9216-859b39aff2ec-kube-api-access-5hsfm\") pod \"nova-metadata-0\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.243137 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.552804 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d11f74a-c4c6-48f7-9c36-29777f70f1a2" path="/var/lib/kubelet/pods/5d11f74a-c4c6-48f7-9c36-29777f70f1a2/volumes" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.554821 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4ef5729-6000-469a-bf66-a53c71bda7e0" path="/var/lib/kubelet/pods/b4ef5729-6000-469a-bf66-a53c71bda7e0/volumes" Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.581607 4894 generic.go:334] "Generic (PLEG): container finished" podID="a57e8fe0-db28-42be-aa0d-f39f751d7897" containerID="02f90fc7a4bd62f67dfd1457046169268063d085c23f8c34c4abcbe11a807ed8" exitCode=0 Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.581673 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pglr9" event={"ID":"a57e8fe0-db28-42be-aa0d-f39f751d7897","Type":"ContainerDied","Data":"02f90fc7a4bd62f67dfd1457046169268063d085c23f8c34c4abcbe11a807ed8"} Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.779495 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2wmmj" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="registry-server" probeResult="failure" output=< Dec 10 19:19:01 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:19:01 crc kubenswrapper[4894]: > Dec 10 19:19:01 crc kubenswrapper[4894]: W1210 19:19:01.798964 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0872c419_8add_4f2c_9216_859b39aff2ec.slice/crio-ea3f428fb0fbd5d736cddaaa1cbbdc176a1f01c804b83de44cd9a95c222d9202 WatchSource:0}: Error finding container ea3f428fb0fbd5d736cddaaa1cbbdc176a1f01c804b83de44cd9a95c222d9202: Status 404 returned error can't find the container with id ea3f428fb0fbd5d736cddaaa1cbbdc176a1f01c804b83de44cd9a95c222d9202 Dec 10 19:19:01 crc kubenswrapper[4894]: I1210 19:19:01.800360 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:02 crc kubenswrapper[4894]: I1210 19:19:02.604188 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0872c419-8add-4f2c-9216-859b39aff2ec","Type":"ContainerStarted","Data":"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677"} Dec 10 19:19:02 crc kubenswrapper[4894]: I1210 19:19:02.604590 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0872c419-8add-4f2c-9216-859b39aff2ec","Type":"ContainerStarted","Data":"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf"} Dec 10 19:19:02 crc kubenswrapper[4894]: I1210 19:19:02.604610 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0872c419-8add-4f2c-9216-859b39aff2ec","Type":"ContainerStarted","Data":"ea3f428fb0fbd5d736cddaaa1cbbdc176a1f01c804b83de44cd9a95c222d9202"} Dec 10 19:19:02 crc kubenswrapper[4894]: I1210 19:19:02.649405 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.649385428 podStartE2EDuration="2.649385428s" podCreationTimestamp="2025-12-10 19:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:02.630959629 +0000 UTC m=+1425.425807451" watchObservedRunningTime="2025-12-10 19:19:02.649385428 +0000 UTC m=+1425.444233230" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.145102 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.245073 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-config-data\") pod \"a57e8fe0-db28-42be-aa0d-f39f751d7897\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.245110 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-combined-ca-bundle\") pod \"a57e8fe0-db28-42be-aa0d-f39f751d7897\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.245156 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-scripts\") pod \"a57e8fe0-db28-42be-aa0d-f39f751d7897\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.245189 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m94rh\" (UniqueName: \"kubernetes.io/projected/a57e8fe0-db28-42be-aa0d-f39f751d7897-kube-api-access-m94rh\") pod \"a57e8fe0-db28-42be-aa0d-f39f751d7897\" (UID: \"a57e8fe0-db28-42be-aa0d-f39f751d7897\") " Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.251514 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57e8fe0-db28-42be-aa0d-f39f751d7897-kube-api-access-m94rh" (OuterVolumeSpecName: "kube-api-access-m94rh") pod "a57e8fe0-db28-42be-aa0d-f39f751d7897" (UID: "a57e8fe0-db28-42be-aa0d-f39f751d7897"). InnerVolumeSpecName "kube-api-access-m94rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.257867 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-scripts" (OuterVolumeSpecName: "scripts") pod "a57e8fe0-db28-42be-aa0d-f39f751d7897" (UID: "a57e8fe0-db28-42be-aa0d-f39f751d7897"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.283884 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-config-data" (OuterVolumeSpecName: "config-data") pod "a57e8fe0-db28-42be-aa0d-f39f751d7897" (UID: "a57e8fe0-db28-42be-aa0d-f39f751d7897"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.314064 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a57e8fe0-db28-42be-aa0d-f39f751d7897" (UID: "a57e8fe0-db28-42be-aa0d-f39f751d7897"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.347341 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.347370 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.347379 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a57e8fe0-db28-42be-aa0d-f39f751d7897-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.347387 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m94rh\" (UniqueName: \"kubernetes.io/projected/a57e8fe0-db28-42be-aa0d-f39f751d7897-kube-api-access-m94rh\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.616144 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-pglr9" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.616144 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-pglr9" event={"ID":"a57e8fe0-db28-42be-aa0d-f39f751d7897","Type":"ContainerDied","Data":"77d9cbdb0a2508e53f6078e52294459cd5cd31cf40ae0b984d0f3b61bf845628"} Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.616208 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77d9cbdb0a2508e53f6078e52294459cd5cd31cf40ae0b984d0f3b61bf845628" Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.825069 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.825366 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-log" containerID="cri-o://effa0119c3772f8bf8b70617cd7d9b4926738127b9f700c404bbba1012e14581" gracePeriod=30 Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.825777 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-api" containerID="cri-o://4953d82118e9f674f5a4f0027f35b08545f0ff14e5451273d1d2db5d670a2bd6" gracePeriod=30 Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.901775 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.902123 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="be401a3a-dd33-46cc-9e88-f402415a996d" containerName="nova-scheduler-scheduler" containerID="cri-o://39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" gracePeriod=30 Dec 10 19:19:03 crc kubenswrapper[4894]: I1210 19:19:03.995368 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:04 crc kubenswrapper[4894]: I1210 19:19:04.634186 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerID="effa0119c3772f8bf8b70617cd7d9b4926738127b9f700c404bbba1012e14581" exitCode=143 Dec 10 19:19:04 crc kubenswrapper[4894]: I1210 19:19:04.634356 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-log" containerID="cri-o://9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf" gracePeriod=30 Dec 10 19:19:04 crc kubenswrapper[4894]: I1210 19:19:04.634591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c","Type":"ContainerDied","Data":"effa0119c3772f8bf8b70617cd7d9b4926738127b9f700c404bbba1012e14581"} Dec 10 19:19:04 crc kubenswrapper[4894]: I1210 19:19:04.634892 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-metadata" containerID="cri-o://6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677" gracePeriod=30 Dec 10 19:19:04 crc kubenswrapper[4894]: E1210 19:19:04.739259 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27 is running failed: container process not found" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:19:04 crc kubenswrapper[4894]: E1210 19:19:04.740238 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27 is running failed: container process not found" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:19:04 crc kubenswrapper[4894]: E1210 19:19:04.740537 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27 is running failed: container process not found" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:19:04 crc kubenswrapper[4894]: E1210 19:19:04.740578 4894 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="be401a3a-dd33-46cc-9e88-f402415a996d" containerName="nova-scheduler-scheduler" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.464276 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.552478 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.562589 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-combined-ca-bundle\") pod \"be401a3a-dd33-46cc-9e88-f402415a996d\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.562645 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-config-data\") pod \"be401a3a-dd33-46cc-9e88-f402415a996d\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.562744 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65qsx\" (UniqueName: \"kubernetes.io/projected/be401a3a-dd33-46cc-9e88-f402415a996d-kube-api-access-65qsx\") pod \"be401a3a-dd33-46cc-9e88-f402415a996d\" (UID: \"be401a3a-dd33-46cc-9e88-f402415a996d\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.569454 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be401a3a-dd33-46cc-9e88-f402415a996d-kube-api-access-65qsx" (OuterVolumeSpecName: "kube-api-access-65qsx") pod "be401a3a-dd33-46cc-9e88-f402415a996d" (UID: "be401a3a-dd33-46cc-9e88-f402415a996d"). InnerVolumeSpecName "kube-api-access-65qsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.610572 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be401a3a-dd33-46cc-9e88-f402415a996d" (UID: "be401a3a-dd33-46cc-9e88-f402415a996d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.615631 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-config-data" (OuterVolumeSpecName: "config-data") pod "be401a3a-dd33-46cc-9e88-f402415a996d" (UID: "be401a3a-dd33-46cc-9e88-f402415a996d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650287 4894 generic.go:334] "Generic (PLEG): container finished" podID="0872c419-8add-4f2c-9216-859b39aff2ec" containerID="6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677" exitCode=0 Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650633 4894 generic.go:334] "Generic (PLEG): container finished" podID="0872c419-8add-4f2c-9216-859b39aff2ec" containerID="9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf" exitCode=143 Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650363 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0872c419-8add-4f2c-9216-859b39aff2ec","Type":"ContainerDied","Data":"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677"} Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650384 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650753 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0872c419-8add-4f2c-9216-859b39aff2ec","Type":"ContainerDied","Data":"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf"} Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650823 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0872c419-8add-4f2c-9216-859b39aff2ec","Type":"ContainerDied","Data":"ea3f428fb0fbd5d736cddaaa1cbbdc176a1f01c804b83de44cd9a95c222d9202"} Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.650883 4894 scope.go:117] "RemoveContainer" containerID="6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.652746 4894 generic.go:334] "Generic (PLEG): container finished" podID="896eec35-0551-4bf5-9afa-179674319fdd" containerID="f60fb0e5527c8ea654ab78a3f1641d06d5e2890ada4ec13c23585124cea12666" exitCode=0 Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.652794 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" event={"ID":"896eec35-0551-4bf5-9afa-179674319fdd","Type":"ContainerDied","Data":"f60fb0e5527c8ea654ab78a3f1641d06d5e2890ada4ec13c23585124cea12666"} Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.664651 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0872c419-8add-4f2c-9216-859b39aff2ec-logs\") pod \"0872c419-8add-4f2c-9216-859b39aff2ec\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.664703 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hsfm\" (UniqueName: \"kubernetes.io/projected/0872c419-8add-4f2c-9216-859b39aff2ec-kube-api-access-5hsfm\") pod \"0872c419-8add-4f2c-9216-859b39aff2ec\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.664771 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-nova-metadata-tls-certs\") pod \"0872c419-8add-4f2c-9216-859b39aff2ec\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.664837 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-combined-ca-bundle\") pod \"0872c419-8add-4f2c-9216-859b39aff2ec\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.664942 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-config-data\") pod \"0872c419-8add-4f2c-9216-859b39aff2ec\" (UID: \"0872c419-8add-4f2c-9216-859b39aff2ec\") " Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.668116 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0872c419-8add-4f2c-9216-859b39aff2ec-logs" (OuterVolumeSpecName: "logs") pod "0872c419-8add-4f2c-9216-859b39aff2ec" (UID: "0872c419-8add-4f2c-9216-859b39aff2ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.668591 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0872c419-8add-4f2c-9216-859b39aff2ec-kube-api-access-5hsfm" (OuterVolumeSpecName: "kube-api-access-5hsfm") pod "0872c419-8add-4f2c-9216-859b39aff2ec" (UID: "0872c419-8add-4f2c-9216-859b39aff2ec"). InnerVolumeSpecName "kube-api-access-5hsfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.670259 4894 generic.go:334] "Generic (PLEG): container finished" podID="be401a3a-dd33-46cc-9e88-f402415a996d" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" exitCode=0 Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.670314 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be401a3a-dd33-46cc-9e88-f402415a996d","Type":"ContainerDied","Data":"39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27"} Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.670333 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.670345 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be401a3a-dd33-46cc-9e88-f402415a996d","Type":"ContainerDied","Data":"a09fb5383ecb24e1b6483575e7005eb12b53c7d8bfa02d7b57111600efc773b2"} Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.673379 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.673407 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be401a3a-dd33-46cc-9e88-f402415a996d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.673417 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65qsx\" (UniqueName: \"kubernetes.io/projected/be401a3a-dd33-46cc-9e88-f402415a996d-kube-api-access-65qsx\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.673428 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0872c419-8add-4f2c-9216-859b39aff2ec-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.673438 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hsfm\" (UniqueName: \"kubernetes.io/projected/0872c419-8add-4f2c-9216-859b39aff2ec-kube-api-access-5hsfm\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.697830 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-config-data" (OuterVolumeSpecName: "config-data") pod "0872c419-8add-4f2c-9216-859b39aff2ec" (UID: "0872c419-8add-4f2c-9216-859b39aff2ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.706927 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0872c419-8add-4f2c-9216-859b39aff2ec" (UID: "0872c419-8add-4f2c-9216-859b39aff2ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.731706 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0872c419-8add-4f2c-9216-859b39aff2ec" (UID: "0872c419-8add-4f2c-9216-859b39aff2ec"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.774873 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.774902 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.774911 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0872c419-8add-4f2c-9216-859b39aff2ec-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.800646 4894 scope.go:117] "RemoveContainer" containerID="9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.810771 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.822149 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.832054 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.832618 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be401a3a-dd33-46cc-9e88-f402415a996d" containerName="nova-scheduler-scheduler" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.832640 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="be401a3a-dd33-46cc-9e88-f402415a996d" containerName="nova-scheduler-scheduler" Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.832658 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-metadata" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.832667 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-metadata" Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.832739 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57e8fe0-db28-42be-aa0d-f39f751d7897" containerName="nova-manage" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.832758 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57e8fe0-db28-42be-aa0d-f39f751d7897" containerName="nova-manage" Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.832791 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-log" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.832812 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-log" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.833109 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-metadata" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.833128 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" containerName="nova-metadata-log" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.833146 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57e8fe0-db28-42be-aa0d-f39f751d7897" containerName="nova-manage" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.833166 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="be401a3a-dd33-46cc-9e88-f402415a996d" containerName="nova-scheduler-scheduler" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.834107 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.840346 4894 scope.go:117] "RemoveContainer" containerID="6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677" Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.841168 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677\": container with ID starting with 6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677 not found: ID does not exist" containerID="6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.841281 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677"} err="failed to get container status \"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677\": rpc error: code = NotFound desc = could not find container \"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677\": container with ID starting with 6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677 not found: ID does not exist" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.841372 4894 scope.go:117] "RemoveContainer" containerID="9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf" Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.841764 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf\": container with ID starting with 9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf not found: ID does not exist" containerID="9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.841787 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf"} err="failed to get container status \"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf\": rpc error: code = NotFound desc = could not find container \"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf\": container with ID starting with 9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf not found: ID does not exist" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.841801 4894 scope.go:117] "RemoveContainer" containerID="6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.849811 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.856247 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677"} err="failed to get container status \"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677\": rpc error: code = NotFound desc = could not find container \"6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677\": container with ID starting with 6dbcf4bf7494c83e127c3c574fd8f938a8d5b7b445cc3583b7e2ab9f082f3677 not found: ID does not exist" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.856323 4894 scope.go:117] "RemoveContainer" containerID="9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.856736 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.875315 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf"} err="failed to get container status \"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf\": rpc error: code = NotFound desc = could not find container \"9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf\": container with ID starting with 9f59ec6456ddcb39cbc7c205ca7c76d8d94684a477d29e156447724fab4be6cf not found: ID does not exist" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.875367 4894 scope.go:117] "RemoveContainer" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.876456 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-config-data\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.876629 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgwjm\" (UniqueName: \"kubernetes.io/projected/19974223-7dae-46bb-a05d-378d0afddec8-kube-api-access-pgwjm\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.876720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.920551 4894 scope.go:117] "RemoveContainer" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" Dec 10 19:19:05 crc kubenswrapper[4894]: E1210 19:19:05.921061 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27\": container with ID starting with 39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27 not found: ID does not exist" containerID="39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.921106 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27"} err="failed to get container status \"39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27\": rpc error: code = NotFound desc = could not find container \"39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27\": container with ID starting with 39e75817e0c37cda60c9dbc8a675eef2984ce236e7b32d190d4923a57fddcf27 not found: ID does not exist" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.982751 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-config-data\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.982841 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgwjm\" (UniqueName: \"kubernetes.io/projected/19974223-7dae-46bb-a05d-378d0afddec8-kube-api-access-pgwjm\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.982896 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.987772 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:05 crc kubenswrapper[4894]: I1210 19:19:05.987959 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-config-data\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.010489 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.010592 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.017045 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgwjm\" (UniqueName: \"kubernetes.io/projected/19974223-7dae-46bb-a05d-378d0afddec8-kube-api-access-pgwjm\") pod \"nova-scheduler-0\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.044408 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.046341 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.054590 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.054880 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.059824 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.160335 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.186830 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.188113 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.188169 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-config-data\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.188265 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9m2j\" (UniqueName: \"kubernetes.io/projected/f498f90b-987c-4be3-9702-5ba84af2e03b-kube-api-access-m9m2j\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.188348 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f498f90b-987c-4be3-9702-5ba84af2e03b-logs\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.289938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.290256 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-config-data\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.290321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9m2j\" (UniqueName: \"kubernetes.io/projected/f498f90b-987c-4be3-9702-5ba84af2e03b-kube-api-access-m9m2j\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.290383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f498f90b-987c-4be3-9702-5ba84af2e03b-logs\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.290444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.291411 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f498f90b-987c-4be3-9702-5ba84af2e03b-logs\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.301963 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-config-data\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.303094 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.305540 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.305843 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9m2j\" (UniqueName: \"kubernetes.io/projected/f498f90b-987c-4be3-9702-5ba84af2e03b-kube-api-access-m9m2j\") pod \"nova-metadata-0\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.411584 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.635404 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.754932 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19974223-7dae-46bb-a05d-378d0afddec8","Type":"ContainerStarted","Data":"c6130adc10c3f0a4ae2495b0de4707339bc56c2d485a1bd21d4559f9683515bc"} Dec 10 19:19:06 crc kubenswrapper[4894]: I1210 19:19:06.984126 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.291271 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.427894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-scripts\") pod \"896eec35-0551-4bf5-9afa-179674319fdd\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.428294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-config-data\") pod \"896eec35-0551-4bf5-9afa-179674319fdd\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.428375 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2czt\" (UniqueName: \"kubernetes.io/projected/896eec35-0551-4bf5-9afa-179674319fdd-kube-api-access-c2czt\") pod \"896eec35-0551-4bf5-9afa-179674319fdd\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.428430 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-combined-ca-bundle\") pod \"896eec35-0551-4bf5-9afa-179674319fdd\" (UID: \"896eec35-0551-4bf5-9afa-179674319fdd\") " Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.432179 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/896eec35-0551-4bf5-9afa-179674319fdd-kube-api-access-c2czt" (OuterVolumeSpecName: "kube-api-access-c2czt") pod "896eec35-0551-4bf5-9afa-179674319fdd" (UID: "896eec35-0551-4bf5-9afa-179674319fdd"). InnerVolumeSpecName "kube-api-access-c2czt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.435138 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-scripts" (OuterVolumeSpecName: "scripts") pod "896eec35-0551-4bf5-9afa-179674319fdd" (UID: "896eec35-0551-4bf5-9afa-179674319fdd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.464074 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "896eec35-0551-4bf5-9afa-179674319fdd" (UID: "896eec35-0551-4bf5-9afa-179674319fdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.466704 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-config-data" (OuterVolumeSpecName: "config-data") pod "896eec35-0551-4bf5-9afa-179674319fdd" (UID: "896eec35-0551-4bf5-9afa-179674319fdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.530779 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2czt\" (UniqueName: \"kubernetes.io/projected/896eec35-0551-4bf5-9afa-179674319fdd-kube-api-access-c2czt\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.530832 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.530870 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.530885 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/896eec35-0551-4bf5-9afa-179674319fdd-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.553124 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0872c419-8add-4f2c-9216-859b39aff2ec" path="/var/lib/kubelet/pods/0872c419-8add-4f2c-9216-859b39aff2ec/volumes" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.553817 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be401a3a-dd33-46cc-9e88-f402415a996d" path="/var/lib/kubelet/pods/be401a3a-dd33-46cc-9e88-f402415a996d/volumes" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.774972 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerID="4953d82118e9f674f5a4f0027f35b08545f0ff14e5451273d1d2db5d670a2bd6" exitCode=0 Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.775050 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c","Type":"ContainerDied","Data":"4953d82118e9f674f5a4f0027f35b08545f0ff14e5451273d1d2db5d670a2bd6"} Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.777151 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" event={"ID":"896eec35-0551-4bf5-9afa-179674319fdd","Type":"ContainerDied","Data":"8de1ccdce0ee7ce42b9bfa842f9d7476a81567edb99d09663e6e3cf58b5fa1dc"} Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.777186 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8de1ccdce0ee7ce42b9bfa842f9d7476a81567edb99d09663e6e3cf58b5fa1dc" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.777158 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8ntxv" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.779356 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19974223-7dae-46bb-a05d-378d0afddec8","Type":"ContainerStarted","Data":"97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4"} Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.790297 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f498f90b-987c-4be3-9702-5ba84af2e03b","Type":"ContainerStarted","Data":"9efc7f1829d23141066e2b433f702a084a92952d196fc63d2553343ba969e812"} Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.790353 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f498f90b-987c-4be3-9702-5ba84af2e03b","Type":"ContainerStarted","Data":"bfe3e77139e89b0df51cc01a877d313db3a25da24ac58b928ea707ce32c7546c"} Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.790368 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f498f90b-987c-4be3-9702-5ba84af2e03b","Type":"ContainerStarted","Data":"71589a9c6075e7779726a11782413df6abdd2fad2baf63762cb20dd71479bede"} Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.807310 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.807289869 podStartE2EDuration="2.807289869s" podCreationTimestamp="2025-12-10 19:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:07.802729486 +0000 UTC m=+1430.597577288" watchObservedRunningTime="2025-12-10 19:19:07.807289869 +0000 UTC m=+1430.602137661" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.846761 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 19:19:07 crc kubenswrapper[4894]: E1210 19:19:07.847286 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="896eec35-0551-4bf5-9afa-179674319fdd" containerName="nova-cell1-conductor-db-sync" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.847300 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="896eec35-0551-4bf5-9afa-179674319fdd" containerName="nova-cell1-conductor-db-sync" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.847504 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="896eec35-0551-4bf5-9afa-179674319fdd" containerName="nova-cell1-conductor-db-sync" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.848314 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.851940 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.872029 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.880347 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.880324397 podStartE2EDuration="2.880324397s" podCreationTimestamp="2025-12-10 19:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:07.830014536 +0000 UTC m=+1430.624862338" watchObservedRunningTime="2025-12-10 19:19:07.880324397 +0000 UTC m=+1430.675172199" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.938327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5404726-0e64-47ba-b50e-07e392a4d737-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.938409 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6btf\" (UniqueName: \"kubernetes.io/projected/b5404726-0e64-47ba-b50e-07e392a4d737-kube-api-access-z6btf\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:07 crc kubenswrapper[4894]: I1210 19:19:07.938435 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5404726-0e64-47ba-b50e-07e392a4d737-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.040081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5404726-0e64-47ba-b50e-07e392a4d737-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.040173 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6btf\" (UniqueName: \"kubernetes.io/projected/b5404726-0e64-47ba-b50e-07e392a4d737-kube-api-access-z6btf\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.040211 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5404726-0e64-47ba-b50e-07e392a4d737-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.044768 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5404726-0e64-47ba-b50e-07e392a4d737-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.046264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5404726-0e64-47ba-b50e-07e392a4d737-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.047711 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.058637 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6btf\" (UniqueName: \"kubernetes.io/projected/b5404726-0e64-47ba-b50e-07e392a4d737-kube-api-access-z6btf\") pod \"nova-cell1-conductor-0\" (UID: \"b5404726-0e64-47ba-b50e-07e392a4d737\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.141772 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-config-data\") pod \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.142221 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsdd6\" (UniqueName: \"kubernetes.io/projected/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-kube-api-access-gsdd6\") pod \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.142394 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-logs\") pod \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.142463 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-combined-ca-bundle\") pod \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\" (UID: \"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c\") " Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.142893 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-logs" (OuterVolumeSpecName: "logs") pod "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" (UID: "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.143044 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.146975 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-kube-api-access-gsdd6" (OuterVolumeSpecName: "kube-api-access-gsdd6") pod "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" (UID: "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c"). InnerVolumeSpecName "kube-api-access-gsdd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.167351 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-config-data" (OuterVolumeSpecName: "config-data") pod "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" (UID: "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.177638 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.207786 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" (UID: "e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.246294 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.246340 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsdd6\" (UniqueName: \"kubernetes.io/projected/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-kube-api-access-gsdd6\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.246355 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.706804 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.801922 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c","Type":"ContainerDied","Data":"c6487005a7ffdc2a9473d4054dc2e6aebb73bb5964098812d4b4917db7807af6"} Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.802395 4894 scope.go:117] "RemoveContainer" containerID="4953d82118e9f674f5a4f0027f35b08545f0ff14e5451273d1d2db5d670a2bd6" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.801955 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.806431 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b5404726-0e64-47ba-b50e-07e392a4d737","Type":"ContainerStarted","Data":"84948642f4d58154597b2a93c35432b5fed6b78a147326d04adae33fd29d6a77"} Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.875372 4894 scope.go:117] "RemoveContainer" containerID="effa0119c3772f8bf8b70617cd7d9b4926738127b9f700c404bbba1012e14581" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.910945 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.931871 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.973294 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:08 crc kubenswrapper[4894]: E1210 19:19:08.974277 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-log" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.974297 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-log" Dec 10 19:19:08 crc kubenswrapper[4894]: E1210 19:19:08.974339 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-api" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.975581 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-api" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.976415 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-api" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.976478 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" containerName="nova-api-log" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.981158 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:08 crc kubenswrapper[4894]: I1210 19:19:08.994250 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.001247 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.065490 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckcjx\" (UniqueName: \"kubernetes.io/projected/25135a0c-3452-4c04-ab19-2757b663b60d-kube-api-access-ckcjx\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.065607 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25135a0c-3452-4c04-ab19-2757b663b60d-logs\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.065774 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.066209 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-config-data\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.167949 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25135a0c-3452-4c04-ab19-2757b663b60d-logs\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.168047 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.168157 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-config-data\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.168219 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckcjx\" (UniqueName: \"kubernetes.io/projected/25135a0c-3452-4c04-ab19-2757b663b60d-kube-api-access-ckcjx\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.168438 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25135a0c-3452-4c04-ab19-2757b663b60d-logs\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.172700 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.173098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-config-data\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.184990 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckcjx\" (UniqueName: \"kubernetes.io/projected/25135a0c-3452-4c04-ab19-2757b663b60d-kube-api-access-ckcjx\") pod \"nova-api-0\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.312482 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.556881 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c" path="/var/lib/kubelet/pods/e5c9e0c4-ff50-4d55-b22a-0fa715a85c6c/volumes" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.783863 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.849830 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b5404726-0e64-47ba-b50e-07e392a4d737","Type":"ContainerStarted","Data":"af9d3cf4cc86b1e00174b755ea74833088cdb56000f7883f1034464dc9fdfd8d"} Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.850964 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.862049 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25135a0c-3452-4c04-ab19-2757b663b60d","Type":"ContainerStarted","Data":"e3efeca71670bd6bf2aaf22ff3c4d847d9da0ba8089c182855719961794c826c"} Dec 10 19:19:09 crc kubenswrapper[4894]: I1210 19:19:09.883995 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.883976757 podStartE2EDuration="2.883976757s" podCreationTimestamp="2025-12-10 19:19:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:09.875307562 +0000 UTC m=+1432.670155354" watchObservedRunningTime="2025-12-10 19:19:09.883976757 +0000 UTC m=+1432.678824549" Dec 10 19:19:10 crc kubenswrapper[4894]: I1210 19:19:10.899565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25135a0c-3452-4c04-ab19-2757b663b60d","Type":"ContainerStarted","Data":"9dc8522cda1ad063e5030ba2ef5bcb39013de24a623ef262dde552978b54cff7"} Dec 10 19:19:10 crc kubenswrapper[4894]: I1210 19:19:10.900076 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25135a0c-3452-4c04-ab19-2757b663b60d","Type":"ContainerStarted","Data":"b88848ed56c7b46619bc821f0f22473f9e7a77df920e533ac5ccbef96d3efda1"} Dec 10 19:19:10 crc kubenswrapper[4894]: I1210 19:19:10.941531 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.94151205 podStartE2EDuration="2.94151205s" podCreationTimestamp="2025-12-10 19:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:10.935215189 +0000 UTC m=+1433.730062981" watchObservedRunningTime="2025-12-10 19:19:10.94151205 +0000 UTC m=+1433.736359842" Dec 10 19:19:11 crc kubenswrapper[4894]: I1210 19:19:11.160697 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 19:19:11 crc kubenswrapper[4894]: I1210 19:19:11.412756 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:19:11 crc kubenswrapper[4894]: I1210 19:19:11.412801 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:19:11 crc kubenswrapper[4894]: I1210 19:19:11.716148 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2wmmj" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="registry-server" probeResult="failure" output=< Dec 10 19:19:11 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:19:11 crc kubenswrapper[4894]: > Dec 10 19:19:15 crc kubenswrapper[4894]: I1210 19:19:15.274170 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 19:19:16 crc kubenswrapper[4894]: I1210 19:19:16.161535 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 19:19:16 crc kubenswrapper[4894]: I1210 19:19:16.192150 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 19:19:16 crc kubenswrapper[4894]: I1210 19:19:16.412395 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:19:16 crc kubenswrapper[4894]: I1210 19:19:16.412471 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:19:16 crc kubenswrapper[4894]: I1210 19:19:16.991616 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 19:19:17 crc kubenswrapper[4894]: I1210 19:19:17.431096 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.215:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:19:17 crc kubenswrapper[4894]: I1210 19:19:17.431121 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.215:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:19:18 crc kubenswrapper[4894]: I1210 19:19:18.205360 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.178517 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.178951 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" containerName="kube-state-metrics" containerID="cri-o://9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2" gracePeriod=30 Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.313375 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.314335 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.741038 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.813422 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf8pt\" (UniqueName: \"kubernetes.io/projected/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d-kube-api-access-bf8pt\") pod \"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d\" (UID: \"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d\") " Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.823143 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d-kube-api-access-bf8pt" (OuterVolumeSpecName: "kube-api-access-bf8pt") pod "5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" (UID: "5fc7a32e-b6e9-4d8c-b007-b7aca357a98d"). InnerVolumeSpecName "kube-api-access-bf8pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:19 crc kubenswrapper[4894]: I1210 19:19:19.915771 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf8pt\" (UniqueName: \"kubernetes.io/projected/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d-kube-api-access-bf8pt\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.015441 4894 generic.go:334] "Generic (PLEG): container finished" podID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" containerID="9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2" exitCode=2 Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.016802 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.016934 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d","Type":"ContainerDied","Data":"9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2"} Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.016994 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5fc7a32e-b6e9-4d8c-b007-b7aca357a98d","Type":"ContainerDied","Data":"fb9d70af946d96dc2705db5b2cd8cd2c46828855bcb074ec498e1d5434e271a8"} Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.017020 4894 scope.go:117] "RemoveContainer" containerID="9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.091153 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.096487 4894 scope.go:117] "RemoveContainer" containerID="9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2" Dec 10 19:19:20 crc kubenswrapper[4894]: E1210 19:19:20.101066 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2\": container with ID starting with 9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2 not found: ID does not exist" containerID="9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.101111 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2"} err="failed to get container status \"9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2\": rpc error: code = NotFound desc = could not find container \"9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2\": container with ID starting with 9b5c3323a501fc60b3d3c5949ad5fe624a1fe309400508ff6df5deccf2cb4df2 not found: ID does not exist" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.109611 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.118504 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:19:20 crc kubenswrapper[4894]: E1210 19:19:20.119032 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" containerName="kube-state-metrics" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.119052 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" containerName="kube-state-metrics" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.119261 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" containerName="kube-state-metrics" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.119977 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.124987 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.125134 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.132445 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.222428 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.222470 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qx2k\" (UniqueName: \"kubernetes.io/projected/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-api-access-4qx2k\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.222504 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.222728 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.325107 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.325602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.325688 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qx2k\" (UniqueName: \"kubernetes.io/projected/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-api-access-4qx2k\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.325772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.329369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.329777 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.330307 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.342509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qx2k\" (UniqueName: \"kubernetes.io/projected/f8a3e9db-6a93-400c-94b2-1d7c5790dc17-kube-api-access-4qx2k\") pod \"kube-state-metrics-0\" (UID: \"f8a3e9db-6a93-400c-94b2-1d7c5790dc17\") " pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.397018 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.217:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.397104 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.217:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.441861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.703367 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:19:20 crc kubenswrapper[4894]: I1210 19:19:20.776383 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.014671 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:19:21 crc kubenswrapper[4894]: W1210 19:19:21.038170 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8a3e9db_6a93_400c_94b2_1d7c5790dc17.slice/crio-0b074548a31e26a0a22768729b42097d20b1b880d10b272aed65edd9039b8c39 WatchSource:0}: Error finding container 0b074548a31e26a0a22768729b42097d20b1b880d10b272aed65edd9039b8c39: Status 404 returned error can't find the container with id 0b074548a31e26a0a22768729b42097d20b1b880d10b272aed65edd9039b8c39 Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.419066 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.419566 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-central-agent" containerID="cri-o://751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d" gracePeriod=30 Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.419677 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="proxy-httpd" containerID="cri-o://d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7" gracePeriod=30 Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.419715 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="sg-core" containerID="cri-o://d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e" gracePeriod=30 Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.419743 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-notification-agent" containerID="cri-o://fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5" gracePeriod=30 Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.441596 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2wmmj"] Dec 10 19:19:21 crc kubenswrapper[4894]: I1210 19:19:21.561718 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fc7a32e-b6e9-4d8c-b007-b7aca357a98d" path="/var/lib/kubelet/pods/5fc7a32e-b6e9-4d8c-b007-b7aca357a98d/volumes" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.039619 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerID="d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7" exitCode=0 Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.040248 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerID="d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e" exitCode=2 Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.040265 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerID="751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d" exitCode=0 Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.039896 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerDied","Data":"d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7"} Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.040382 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerDied","Data":"d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e"} Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.040403 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerDied","Data":"751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d"} Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.042188 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f8a3e9db-6a93-400c-94b2-1d7c5790dc17","Type":"ContainerStarted","Data":"7709ec206a3584e67d634f5586e2301659189a3ac7200c2fecfed106eabf82a4"} Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.042249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f8a3e9db-6a93-400c-94b2-1d7c5790dc17","Type":"ContainerStarted","Data":"0b074548a31e26a0a22768729b42097d20b1b880d10b272aed65edd9039b8c39"} Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.042317 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2wmmj" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="registry-server" containerID="cri-o://7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8" gracePeriod=2 Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.061988 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.6704394169999999 podStartE2EDuration="2.061971068s" podCreationTimestamp="2025-12-10 19:19:20 +0000 UTC" firstStartedPulling="2025-12-10 19:19:21.040182243 +0000 UTC m=+1443.835030035" lastFinishedPulling="2025-12-10 19:19:21.431713884 +0000 UTC m=+1444.226561686" observedRunningTime="2025-12-10 19:19:22.054921478 +0000 UTC m=+1444.849769290" watchObservedRunningTime="2025-12-10 19:19:22.061971068 +0000 UTC m=+1444.856818870" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.627097 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.691269 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-catalog-content\") pod \"74a13406-618a-4f69-b9de-fafec8c37346\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.691614 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-utilities\") pod \"74a13406-618a-4f69-b9de-fafec8c37346\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.691789 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cnmv\" (UniqueName: \"kubernetes.io/projected/74a13406-618a-4f69-b9de-fafec8c37346-kube-api-access-7cnmv\") pod \"74a13406-618a-4f69-b9de-fafec8c37346\" (UID: \"74a13406-618a-4f69-b9de-fafec8c37346\") " Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.692317 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-utilities" (OuterVolumeSpecName: "utilities") pod "74a13406-618a-4f69-b9de-fafec8c37346" (UID: "74a13406-618a-4f69-b9de-fafec8c37346"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.698228 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a13406-618a-4f69-b9de-fafec8c37346-kube-api-access-7cnmv" (OuterVolumeSpecName: "kube-api-access-7cnmv") pod "74a13406-618a-4f69-b9de-fafec8c37346" (UID: "74a13406-618a-4f69-b9de-fafec8c37346"). InnerVolumeSpecName "kube-api-access-7cnmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.790556 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74a13406-618a-4f69-b9de-fafec8c37346" (UID: "74a13406-618a-4f69-b9de-fafec8c37346"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.794506 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.794532 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cnmv\" (UniqueName: \"kubernetes.io/projected/74a13406-618a-4f69-b9de-fafec8c37346-kube-api-access-7cnmv\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:22 crc kubenswrapper[4894]: I1210 19:19:22.794541 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a13406-618a-4f69-b9de-fafec8c37346-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.053360 4894 generic.go:334] "Generic (PLEG): container finished" podID="74a13406-618a-4f69-b9de-fafec8c37346" containerID="7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8" exitCode=0 Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.053618 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2wmmj" event={"ID":"74a13406-618a-4f69-b9de-fafec8c37346","Type":"ContainerDied","Data":"7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8"} Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.053719 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2wmmj" event={"ID":"74a13406-618a-4f69-b9de-fafec8c37346","Type":"ContainerDied","Data":"aaf19152b19f8165c3194dbb89064ed8def1bec63ea36ea5b7ce5032bcf100fc"} Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.053744 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.053771 4894 scope.go:117] "RemoveContainer" containerID="7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.053908 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2wmmj" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.078471 4894 scope.go:117] "RemoveContainer" containerID="b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.095673 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2wmmj"] Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.107781 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2wmmj"] Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.125123 4894 scope.go:117] "RemoveContainer" containerID="0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.410028 4894 scope.go:117] "RemoveContainer" containerID="7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8" Dec 10 19:19:23 crc kubenswrapper[4894]: E1210 19:19:23.413911 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8\": container with ID starting with 7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8 not found: ID does not exist" containerID="7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.413952 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8"} err="failed to get container status \"7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8\": rpc error: code = NotFound desc = could not find container \"7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8\": container with ID starting with 7b5b3161dba738fb3d6f190ba00f8f331fc3f3d0161c122dedcb64a306db9bd8 not found: ID does not exist" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.413974 4894 scope.go:117] "RemoveContainer" containerID="b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd" Dec 10 19:19:23 crc kubenswrapper[4894]: E1210 19:19:23.414625 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd\": container with ID starting with b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd not found: ID does not exist" containerID="b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.414677 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd"} err="failed to get container status \"b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd\": rpc error: code = NotFound desc = could not find container \"b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd\": container with ID starting with b94fea5ac6d7bd6b81865f2b1e590482f019640ee2bb27cf5def2538769883cd not found: ID does not exist" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.414706 4894 scope.go:117] "RemoveContainer" containerID="0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3" Dec 10 19:19:23 crc kubenswrapper[4894]: E1210 19:19:23.415099 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3\": container with ID starting with 0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3 not found: ID does not exist" containerID="0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.415122 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3"} err="failed to get container status \"0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3\": rpc error: code = NotFound desc = could not find container \"0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3\": container with ID starting with 0136d3ba4599634cd6598e90614f969890b06c26149512904c4ba707c59015e3 not found: ID does not exist" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.506304 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.556988 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a13406-618a-4f69-b9de-fafec8c37346" path="/var/lib/kubelet/pods/74a13406-618a-4f69-b9de-fafec8c37346/volumes" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617071 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-config-data\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617107 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-combined-ca-bundle\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617206 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-log-httpd\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617243 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-sg-core-conf-yaml\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617293 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-scripts\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617334 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-949s8\" (UniqueName: \"kubernetes.io/projected/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-kube-api-access-949s8\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617405 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-run-httpd\") pod \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\" (UID: \"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e\") " Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.617995 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.618141 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.618149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.629023 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-scripts" (OuterVolumeSpecName: "scripts") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.638879 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-kube-api-access-949s8" (OuterVolumeSpecName: "kube-api-access-949s8") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "kube-api-access-949s8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.656930 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.695104 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.719919 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.719951 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.719964 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-949s8\" (UniqueName: \"kubernetes.io/projected/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-kube-api-access-949s8\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.719976 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.719986 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.736552 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-config-data" (OuterVolumeSpecName: "config-data") pod "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" (UID: "5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4894]: I1210 19:19:23.821992 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.068686 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerID="fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5" exitCode=0 Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.068752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerDied","Data":"fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5"} Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.068825 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.069729 4894 scope.go:117] "RemoveContainer" containerID="d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.069894 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e","Type":"ContainerDied","Data":"f4d569d4ad1937809089642c84bc548ae90ec23332cc57bda552d7ca80df370c"} Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.101117 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.109925 4894 scope.go:117] "RemoveContainer" containerID="d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.122515 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.140096 4894 scope.go:117] "RemoveContainer" containerID="fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.149427 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.149921 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="sg-core" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.149940 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="sg-core" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.149967 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="proxy-httpd" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.149976 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="proxy-httpd" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.149999 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="extract-utilities" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150007 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="extract-utilities" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.150026 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-central-agent" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150034 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-central-agent" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.150043 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="extract-content" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150051 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="extract-content" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.150062 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="registry-server" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150072 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="registry-server" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.150084 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-notification-agent" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150092 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-notification-agent" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150326 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-central-agent" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150345 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a13406-618a-4f69-b9de-fafec8c37346" containerName="registry-server" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150371 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="ceilometer-notification-agent" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150387 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="proxy-httpd" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.150403 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" containerName="sg-core" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.152688 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.157186 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.157552 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.157896 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.173776 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.189874 4894 scope.go:117] "RemoveContainer" containerID="751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.218238 4894 scope.go:117] "RemoveContainer" containerID="d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.218690 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7\": container with ID starting with d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7 not found: ID does not exist" containerID="d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.218718 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7"} err="failed to get container status \"d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7\": rpc error: code = NotFound desc = could not find container \"d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7\": container with ID starting with d0d1e42f9c5fd683f55c6b0e1750f1dfd884eae437fc681fd3fcaab80385bec7 not found: ID does not exist" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.218740 4894 scope.go:117] "RemoveContainer" containerID="d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.219113 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e\": container with ID starting with d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e not found: ID does not exist" containerID="d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.219208 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e"} err="failed to get container status \"d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e\": rpc error: code = NotFound desc = could not find container \"d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e\": container with ID starting with d6a82ff7f2590f553467be0ce530584f009e17fed07c09b34ad58cb109d4172e not found: ID does not exist" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.219294 4894 scope.go:117] "RemoveContainer" containerID="fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.219632 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5\": container with ID starting with fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5 not found: ID does not exist" containerID="fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.219655 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5"} err="failed to get container status \"fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5\": rpc error: code = NotFound desc = could not find container \"fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5\": container with ID starting with fdb024575a0bb052e7c32d52a45a33b48a32b879d545594c17695cb5e659aaa5 not found: ID does not exist" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.219669 4894 scope.go:117] "RemoveContainer" containerID="751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d" Dec 10 19:19:24 crc kubenswrapper[4894]: E1210 19:19:24.220083 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d\": container with ID starting with 751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d not found: ID does not exist" containerID="751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.220105 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d"} err="failed to get container status \"751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d\": rpc error: code = NotFound desc = could not find container \"751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d\": container with ID starting with 751ff2aebba52bbfaa4f85e6d195ea7e6b6e7000849859d3b57f8d854e5bf68d not found: ID does not exist" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.233552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqjkl\" (UniqueName: \"kubernetes.io/projected/8ae00430-da8a-4541-a90d-7d7031bd6e38-kube-api-access-dqjkl\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.233746 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-config-data\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.233878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.233998 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.234198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-scripts\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.234284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.234385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-run-httpd\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.234470 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-log-httpd\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336554 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-scripts\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336596 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336621 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-run-httpd\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336639 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-log-httpd\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336710 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqjkl\" (UniqueName: \"kubernetes.io/projected/8ae00430-da8a-4541-a90d-7d7031bd6e38-kube-api-access-dqjkl\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336727 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-config-data\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.336806 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.337759 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-log-httpd\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.338061 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-run-httpd\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.340362 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-scripts\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.341775 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.342040 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-config-data\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.343345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.344311 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.355282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqjkl\" (UniqueName: \"kubernetes.io/projected/8ae00430-da8a-4541-a90d-7d7031bd6e38-kube-api-access-dqjkl\") pod \"ceilometer-0\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.488669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:24 crc kubenswrapper[4894]: I1210 19:19:24.972288 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:25 crc kubenswrapper[4894]: I1210 19:19:25.085165 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerStarted","Data":"693c8a5fd7eb14cf83cd048ac1d3a5dd74034eea6bdbf0043086c00e97d29d53"} Dec 10 19:19:25 crc kubenswrapper[4894]: I1210 19:19:25.554607 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e" path="/var/lib/kubelet/pods/5ef64c2d-c03b-49f0-b7aa-2a3eec2f550e/volumes" Dec 10 19:19:26 crc kubenswrapper[4894]: I1210 19:19:26.099150 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerStarted","Data":"971e9aec8960e5d957003d9a250573cbc04c1737d027e426941ae6012505f59c"} Dec 10 19:19:26 crc kubenswrapper[4894]: I1210 19:19:26.422008 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:19:26 crc kubenswrapper[4894]: I1210 19:19:26.423216 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:19:26 crc kubenswrapper[4894]: I1210 19:19:26.435653 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.089754 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.111704 4894 generic.go:334] "Generic (PLEG): container finished" podID="02f732ef-d8c7-437b-8079-e6da9a0537e9" containerID="378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2" exitCode=137 Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.111780 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"02f732ef-d8c7-437b-8079-e6da9a0537e9","Type":"ContainerDied","Data":"378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2"} Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.111874 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"02f732ef-d8c7-437b-8079-e6da9a0537e9","Type":"ContainerDied","Data":"6f5e9c01e8daead9f91b78c43360b20f3bf3ced80d0add9c4d6123b6d020e35c"} Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.111791 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.111909 4894 scope.go:117] "RemoveContainer" containerID="378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.119930 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerStarted","Data":"1e581844ae6451ac6f0d2882893a2cb4989a19a4b8ec2b4611ea4c0019d2fa37"} Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.144916 4894 scope.go:117] "RemoveContainer" containerID="378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.150280 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:19:27 crc kubenswrapper[4894]: E1210 19:19:27.153676 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2\": container with ID starting with 378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2 not found: ID does not exist" containerID="378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.153713 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2"} err="failed to get container status \"378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2\": rpc error: code = NotFound desc = could not find container \"378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2\": container with ID starting with 378061314fe88fefcf2f4b37c24ee1253428150711599d034e1f356616f072f2 not found: ID does not exist" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.204613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-config-data\") pod \"02f732ef-d8c7-437b-8079-e6da9a0537e9\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.204835 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56kz8\" (UniqueName: \"kubernetes.io/projected/02f732ef-d8c7-437b-8079-e6da9a0537e9-kube-api-access-56kz8\") pod \"02f732ef-d8c7-437b-8079-e6da9a0537e9\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.205013 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-combined-ca-bundle\") pod \"02f732ef-d8c7-437b-8079-e6da9a0537e9\" (UID: \"02f732ef-d8c7-437b-8079-e6da9a0537e9\") " Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.217828 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f732ef-d8c7-437b-8079-e6da9a0537e9-kube-api-access-56kz8" (OuterVolumeSpecName: "kube-api-access-56kz8") pod "02f732ef-d8c7-437b-8079-e6da9a0537e9" (UID: "02f732ef-d8c7-437b-8079-e6da9a0537e9"). InnerVolumeSpecName "kube-api-access-56kz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.239750 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02f732ef-d8c7-437b-8079-e6da9a0537e9" (UID: "02f732ef-d8c7-437b-8079-e6da9a0537e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.257991 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-config-data" (OuterVolumeSpecName: "config-data") pod "02f732ef-d8c7-437b-8079-e6da9a0537e9" (UID: "02f732ef-d8c7-437b-8079-e6da9a0537e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.307586 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56kz8\" (UniqueName: \"kubernetes.io/projected/02f732ef-d8c7-437b-8079-e6da9a0537e9-kube-api-access-56kz8\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.307619 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.307633 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f732ef-d8c7-437b-8079-e6da9a0537e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.449273 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.456571 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.474211 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:19:27 crc kubenswrapper[4894]: E1210 19:19:27.474642 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f732ef-d8c7-437b-8079-e6da9a0537e9" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.474660 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f732ef-d8c7-437b-8079-e6da9a0537e9" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.474919 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f732ef-d8c7-437b-8079-e6da9a0537e9" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.475596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.480494 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.480596 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.480742 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.498353 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.514284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.514553 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.514576 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.514638 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj65p\" (UniqueName: \"kubernetes.io/projected/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-kube-api-access-gj65p\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.514880 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.566753 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f732ef-d8c7-437b-8079-e6da9a0537e9" path="/var/lib/kubelet/pods/02f732ef-d8c7-437b-8079-e6da9a0537e9/volumes" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.616295 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.616376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.616433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.616452 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.616494 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj65p\" (UniqueName: \"kubernetes.io/projected/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-kube-api-access-gj65p\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.624932 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.625488 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.627191 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.637357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.652501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj65p\" (UniqueName: \"kubernetes.io/projected/921bff93-5bc2-41c2-9d44-eac7ccdc50c1-kube-api-access-gj65p\") pod \"nova-cell1-novncproxy-0\" (UID: \"921bff93-5bc2-41c2-9d44-eac7ccdc50c1\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:27 crc kubenswrapper[4894]: I1210 19:19:27.790448 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:28 crc kubenswrapper[4894]: I1210 19:19:28.138153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerStarted","Data":"b81fe73086f0b57f90e1adab0c75192bb8be18ac900a0ff79deff4092b034e33"} Dec 10 19:19:28 crc kubenswrapper[4894]: I1210 19:19:28.328130 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:19:28 crc kubenswrapper[4894]: W1210 19:19:28.338265 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod921bff93_5bc2_41c2_9d44_eac7ccdc50c1.slice/crio-c94241b01849939fb6e2159f51115957df5c8232a810616a73de0714b15318b8 WatchSource:0}: Error finding container c94241b01849939fb6e2159f51115957df5c8232a810616a73de0714b15318b8: Status 404 returned error can't find the container with id c94241b01849939fb6e2159f51115957df5c8232a810616a73de0714b15318b8 Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.151336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerStarted","Data":"cf10898a75cb3713603bc133681e6b964160ca7b45f1b33038c45b2417290d4f"} Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.151709 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.156945 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"921bff93-5bc2-41c2-9d44-eac7ccdc50c1","Type":"ContainerStarted","Data":"022671bf7639547c45c410e63dbd3951aed9e2d3df49e71717a0bfde86b00081"} Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.157185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"921bff93-5bc2-41c2-9d44-eac7ccdc50c1","Type":"ContainerStarted","Data":"c94241b01849939fb6e2159f51115957df5c8232a810616a73de0714b15318b8"} Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.184303 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.694998064 podStartE2EDuration="5.184284897s" podCreationTimestamp="2025-12-10 19:19:24 +0000 UTC" firstStartedPulling="2025-12-10 19:19:24.969331377 +0000 UTC m=+1447.764179169" lastFinishedPulling="2025-12-10 19:19:28.45861821 +0000 UTC m=+1451.253466002" observedRunningTime="2025-12-10 19:19:29.174898723 +0000 UTC m=+1451.969746525" watchObservedRunningTime="2025-12-10 19:19:29.184284897 +0000 UTC m=+1451.979132689" Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.213753 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.213733604 podStartE2EDuration="2.213733604s" podCreationTimestamp="2025-12-10 19:19:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:29.203558919 +0000 UTC m=+1451.998406731" watchObservedRunningTime="2025-12-10 19:19:29.213733604 +0000 UTC m=+1452.008581396" Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.316011 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.316727 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.316946 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:19:29 crc kubenswrapper[4894]: I1210 19:19:29.319193 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.167344 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.174041 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.394409 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78468d7767-8ppb5"] Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.396275 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.418187 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78468d7767-8ppb5"] Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.453662 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.587625 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-svc\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.587692 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6r9z\" (UniqueName: \"kubernetes.io/projected/f46b5f26-4bcb-4796-9a8b-706735947bc9-kube-api-access-f6r9z\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.587760 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-swift-storage-0\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.587867 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-config\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.587915 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-sb\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.588136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-nb\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.690241 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-sb\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.690355 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-nb\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.690505 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-svc\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.690551 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6r9z\" (UniqueName: \"kubernetes.io/projected/f46b5f26-4bcb-4796-9a8b-706735947bc9-kube-api-access-f6r9z\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.690608 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-swift-storage-0\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.690702 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-config\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.691200 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-sb\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.691274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-nb\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.691394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-config\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.691517 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-svc\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.691578 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-swift-storage-0\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.712616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6r9z\" (UniqueName: \"kubernetes.io/projected/f46b5f26-4bcb-4796-9a8b-706735947bc9-kube-api-access-f6r9z\") pod \"dnsmasq-dns-78468d7767-8ppb5\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:30 crc kubenswrapper[4894]: I1210 19:19:30.721574 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:31 crc kubenswrapper[4894]: I1210 19:19:31.304753 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78468d7767-8ppb5"] Dec 10 19:19:31 crc kubenswrapper[4894]: W1210 19:19:31.313344 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf46b5f26_4bcb_4796_9a8b_706735947bc9.slice/crio-a651524e520c39506bce38ecd09a1c0f7119bc9c7b3c4e93dd6425d0e64e80a5 WatchSource:0}: Error finding container a651524e520c39506bce38ecd09a1c0f7119bc9c7b3c4e93dd6425d0e64e80a5: Status 404 returned error can't find the container with id a651524e520c39506bce38ecd09a1c0f7119bc9c7b3c4e93dd6425d0e64e80a5 Dec 10 19:19:32 crc kubenswrapper[4894]: I1210 19:19:32.221802 4894 generic.go:334] "Generic (PLEG): container finished" podID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerID="d194c2132d2625084689fc37912f640205a1e5d3f84e476b25ba84dcfd211378" exitCode=0 Dec 10 19:19:32 crc kubenswrapper[4894]: I1210 19:19:32.221929 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" event={"ID":"f46b5f26-4bcb-4796-9a8b-706735947bc9","Type":"ContainerDied","Data":"d194c2132d2625084689fc37912f640205a1e5d3f84e476b25ba84dcfd211378"} Dec 10 19:19:32 crc kubenswrapper[4894]: I1210 19:19:32.222191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" event={"ID":"f46b5f26-4bcb-4796-9a8b-706735947bc9","Type":"ContainerStarted","Data":"a651524e520c39506bce38ecd09a1c0f7119bc9c7b3c4e93dd6425d0e64e80a5"} Dec 10 19:19:32 crc kubenswrapper[4894]: I1210 19:19:32.790982 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:33 crc kubenswrapper[4894]: I1210 19:19:33.232133 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" event={"ID":"f46b5f26-4bcb-4796-9a8b-706735947bc9","Type":"ContainerStarted","Data":"3742f7be0244d31b9994feb14d1b92fc6e804b562b8507d90ebbb61a2a977c8e"} Dec 10 19:19:33 crc kubenswrapper[4894]: I1210 19:19:33.233575 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:33 crc kubenswrapper[4894]: I1210 19:19:33.278404 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" podStartSLOduration=3.278383306 podStartE2EDuration="3.278383306s" podCreationTimestamp="2025-12-10 19:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:33.269572107 +0000 UTC m=+1456.064419909" watchObservedRunningTime="2025-12-10 19:19:33.278383306 +0000 UTC m=+1456.073231098" Dec 10 19:19:33 crc kubenswrapper[4894]: I1210 19:19:33.701565 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:33 crc kubenswrapper[4894]: I1210 19:19:33.701773 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-log" containerID="cri-o://b88848ed56c7b46619bc821f0f22473f9e7a77df920e533ac5ccbef96d3efda1" gracePeriod=30 Dec 10 19:19:33 crc kubenswrapper[4894]: I1210 19:19:33.702255 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-api" containerID="cri-o://9dc8522cda1ad063e5030ba2ef5bcb39013de24a623ef262dde552978b54cff7" gracePeriod=30 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.024298 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.024905 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-central-agent" containerID="cri-o://971e9aec8960e5d957003d9a250573cbc04c1737d027e426941ae6012505f59c" gracePeriod=30 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.024933 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="proxy-httpd" containerID="cri-o://cf10898a75cb3713603bc133681e6b964160ca7b45f1b33038c45b2417290d4f" gracePeriod=30 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.025037 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-notification-agent" containerID="cri-o://1e581844ae6451ac6f0d2882893a2cb4989a19a4b8ec2b4611ea4c0019d2fa37" gracePeriod=30 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.025089 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="sg-core" containerID="cri-o://b81fe73086f0b57f90e1adab0c75192bb8be18ac900a0ff79deff4092b034e33" gracePeriod=30 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.255082 4894 generic.go:334] "Generic (PLEG): container finished" podID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerID="cf10898a75cb3713603bc133681e6b964160ca7b45f1b33038c45b2417290d4f" exitCode=0 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.255128 4894 generic.go:334] "Generic (PLEG): container finished" podID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerID="b81fe73086f0b57f90e1adab0c75192bb8be18ac900a0ff79deff4092b034e33" exitCode=2 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.255184 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerDied","Data":"cf10898a75cb3713603bc133681e6b964160ca7b45f1b33038c45b2417290d4f"} Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.255216 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerDied","Data":"b81fe73086f0b57f90e1adab0c75192bb8be18ac900a0ff79deff4092b034e33"} Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.257548 4894 generic.go:334] "Generic (PLEG): container finished" podID="25135a0c-3452-4c04-ab19-2757b663b60d" containerID="b88848ed56c7b46619bc821f0f22473f9e7a77df920e533ac5ccbef96d3efda1" exitCode=143 Dec 10 19:19:34 crc kubenswrapper[4894]: I1210 19:19:34.257856 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25135a0c-3452-4c04-ab19-2757b663b60d","Type":"ContainerDied","Data":"b88848ed56c7b46619bc821f0f22473f9e7a77df920e533ac5ccbef96d3efda1"} Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.276372 4894 generic.go:334] "Generic (PLEG): container finished" podID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerID="1e581844ae6451ac6f0d2882893a2cb4989a19a4b8ec2b4611ea4c0019d2fa37" exitCode=0 Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.276724 4894 generic.go:334] "Generic (PLEG): container finished" podID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerID="971e9aec8960e5d957003d9a250573cbc04c1737d027e426941ae6012505f59c" exitCode=0 Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.276461 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerDied","Data":"1e581844ae6451ac6f0d2882893a2cb4989a19a4b8ec2b4611ea4c0019d2fa37"} Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.276854 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerDied","Data":"971e9aec8960e5d957003d9a250573cbc04c1737d027e426941ae6012505f59c"} Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.696629 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818488 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-scripts\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818570 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-sg-core-conf-yaml\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818617 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-config-data\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818642 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-log-httpd\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818670 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-combined-ca-bundle\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818764 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-run-httpd\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818866 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqjkl\" (UniqueName: \"kubernetes.io/projected/8ae00430-da8a-4541-a90d-7d7031bd6e38-kube-api-access-dqjkl\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.818897 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-ceilometer-tls-certs\") pod \"8ae00430-da8a-4541-a90d-7d7031bd6e38\" (UID: \"8ae00430-da8a-4541-a90d-7d7031bd6e38\") " Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.820461 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.820958 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.824964 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-scripts" (OuterVolumeSpecName: "scripts") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.826326 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae00430-da8a-4541-a90d-7d7031bd6e38-kube-api-access-dqjkl" (OuterVolumeSpecName: "kube-api-access-dqjkl") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "kube-api-access-dqjkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.866061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.887908 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.921041 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqjkl\" (UniqueName: \"kubernetes.io/projected/8ae00430-da8a-4541-a90d-7d7031bd6e38-kube-api-access-dqjkl\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.921076 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.921084 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.921094 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.921102 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.921111 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8ae00430-da8a-4541-a90d-7d7031bd6e38-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.924155 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:35 crc kubenswrapper[4894]: I1210 19:19:35.940718 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-config-data" (OuterVolumeSpecName: "config-data") pod "8ae00430-da8a-4541-a90d-7d7031bd6e38" (UID: "8ae00430-da8a-4541-a90d-7d7031bd6e38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.023311 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.023344 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ae00430-da8a-4541-a90d-7d7031bd6e38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.289148 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8ae00430-da8a-4541-a90d-7d7031bd6e38","Type":"ContainerDied","Data":"693c8a5fd7eb14cf83cd048ac1d3a5dd74034eea6bdbf0043086c00e97d29d53"} Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.289214 4894 scope.go:117] "RemoveContainer" containerID="cf10898a75cb3713603bc133681e6b964160ca7b45f1b33038c45b2417290d4f" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.289231 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.310492 4894 scope.go:117] "RemoveContainer" containerID="b81fe73086f0b57f90e1adab0c75192bb8be18ac900a0ff79deff4092b034e33" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.333007 4894 scope.go:117] "RemoveContainer" containerID="1e581844ae6451ac6f0d2882893a2cb4989a19a4b8ec2b4611ea4c0019d2fa37" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.334121 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.346260 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.359505 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:36 crc kubenswrapper[4894]: E1210 19:19:36.360226 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="proxy-httpd" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360271 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="proxy-httpd" Dec 10 19:19:36 crc kubenswrapper[4894]: E1210 19:19:36.360310 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="sg-core" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360319 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="sg-core" Dec 10 19:19:36 crc kubenswrapper[4894]: E1210 19:19:36.360362 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-central-agent" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360374 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-central-agent" Dec 10 19:19:36 crc kubenswrapper[4894]: E1210 19:19:36.360385 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-notification-agent" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360390 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-notification-agent" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360682 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="sg-core" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360711 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-notification-agent" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360725 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="proxy-httpd" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.360746 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" containerName="ceilometer-central-agent" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.361932 4894 scope.go:117] "RemoveContainer" containerID="971e9aec8960e5d957003d9a250573cbc04c1737d027e426941ae6012505f59c" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.362809 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.364618 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.369298 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.369521 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.376597 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433038 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-log-httpd\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-run-httpd\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433233 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-scripts\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433260 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbpnj\" (UniqueName: \"kubernetes.io/projected/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-kube-api-access-cbpnj\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433318 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433348 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.433390 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-config-data\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535039 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-log-httpd\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535176 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-run-httpd\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535309 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-scripts\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbpnj\" (UniqueName: \"kubernetes.io/projected/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-kube-api-access-cbpnj\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535421 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535465 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.535501 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-config-data\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.536214 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-log-httpd\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.536355 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-run-httpd\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.539329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.540017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.540055 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-config-data\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.540578 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-scripts\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.551742 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.566919 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbpnj\" (UniqueName: \"kubernetes.io/projected/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-kube-api-access-cbpnj\") pod \"ceilometer-0\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " pod="openstack/ceilometer-0" Dec 10 19:19:36 crc kubenswrapper[4894]: I1210 19:19:36.690927 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.309210 4894 generic.go:334] "Generic (PLEG): container finished" podID="25135a0c-3452-4c04-ab19-2757b663b60d" containerID="9dc8522cda1ad063e5030ba2ef5bcb39013de24a623ef262dde552978b54cff7" exitCode=0 Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.309266 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25135a0c-3452-4c04-ab19-2757b663b60d","Type":"ContainerDied","Data":"9dc8522cda1ad063e5030ba2ef5bcb39013de24a623ef262dde552978b54cff7"} Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.398078 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.597135 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ae00430-da8a-4541-a90d-7d7031bd6e38" path="/var/lib/kubelet/pods/8ae00430-da8a-4541-a90d-7d7031bd6e38/volumes" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.663153 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.704211 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckcjx\" (UniqueName: \"kubernetes.io/projected/25135a0c-3452-4c04-ab19-2757b663b60d-kube-api-access-ckcjx\") pod \"25135a0c-3452-4c04-ab19-2757b663b60d\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.705133 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25135a0c-3452-4c04-ab19-2757b663b60d-logs\") pod \"25135a0c-3452-4c04-ab19-2757b663b60d\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.705338 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-config-data\") pod \"25135a0c-3452-4c04-ab19-2757b663b60d\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.708137 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25135a0c-3452-4c04-ab19-2757b663b60d-logs" (OuterVolumeSpecName: "logs") pod "25135a0c-3452-4c04-ab19-2757b663b60d" (UID: "25135a0c-3452-4c04-ab19-2757b663b60d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.721510 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25135a0c-3452-4c04-ab19-2757b663b60d-kube-api-access-ckcjx" (OuterVolumeSpecName: "kube-api-access-ckcjx") pod "25135a0c-3452-4c04-ab19-2757b663b60d" (UID: "25135a0c-3452-4c04-ab19-2757b663b60d"). InnerVolumeSpecName "kube-api-access-ckcjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.746142 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-config-data" (OuterVolumeSpecName: "config-data") pod "25135a0c-3452-4c04-ab19-2757b663b60d" (UID: "25135a0c-3452-4c04-ab19-2757b663b60d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.791175 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.809037 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-combined-ca-bundle\") pod \"25135a0c-3452-4c04-ab19-2757b663b60d\" (UID: \"25135a0c-3452-4c04-ab19-2757b663b60d\") " Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.809791 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckcjx\" (UniqueName: \"kubernetes.io/projected/25135a0c-3452-4c04-ab19-2757b663b60d-kube-api-access-ckcjx\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.809830 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25135a0c-3452-4c04-ab19-2757b663b60d-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.809864 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.821353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.870592 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25135a0c-3452-4c04-ab19-2757b663b60d" (UID: "25135a0c-3452-4c04-ab19-2757b663b60d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:37 crc kubenswrapper[4894]: I1210 19:19:37.910731 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25135a0c-3452-4c04-ab19-2757b663b60d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.325310 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerStarted","Data":"3729d19ec9376b06cac616f59a5089e344b9b3bd31c4c0a9d3637082a47b8f13"} Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.325356 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerStarted","Data":"a7b2b495f8d429b02bce7ee638dff34c6fbf9082b44e7c64f465670747bd1c1e"} Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.330726 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.330741 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25135a0c-3452-4c04-ab19-2757b663b60d","Type":"ContainerDied","Data":"e3efeca71670bd6bf2aaf22ff3c4d847d9da0ba8089c182855719961794c826c"} Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.330806 4894 scope.go:117] "RemoveContainer" containerID="9dc8522cda1ad063e5030ba2ef5bcb39013de24a623ef262dde552978b54cff7" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.349239 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.367001 4894 scope.go:117] "RemoveContainer" containerID="b88848ed56c7b46619bc821f0f22473f9e7a77df920e533ac5ccbef96d3efda1" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.440908 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.453289 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.467162 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:38 crc kubenswrapper[4894]: E1210 19:19:38.467771 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-log" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.467927 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-log" Dec 10 19:19:38 crc kubenswrapper[4894]: E1210 19:19:38.467989 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-api" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.468041 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-api" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.468313 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-api" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.468402 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" containerName="nova-api-log" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.469730 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.477639 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.477911 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.478063 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.487487 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.647394 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.647962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee0c839-adff-484f-b5f5-1a836ead634f-logs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.648100 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.648228 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l64c7\" (UniqueName: \"kubernetes.io/projected/5ee0c839-adff-484f-b5f5-1a836ead634f-kube-api-access-l64c7\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.648319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-config-data\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.648450 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-public-tls-certs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.699990 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9pwpt"] Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.701838 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.710152 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.710299 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.714150 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9pwpt"] Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753312 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-config-data\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753371 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-public-tls-certs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-config-data\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753474 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee0c839-adff-484f-b5f5-1a836ead634f-logs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-scripts\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753535 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753570 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrtmf\" (UniqueName: \"kubernetes.io/projected/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-kube-api-access-wrtmf\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.753616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l64c7\" (UniqueName: \"kubernetes.io/projected/5ee0c839-adff-484f-b5f5-1a836ead634f-kube-api-access-l64c7\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.754484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee0c839-adff-484f-b5f5-1a836ead634f-logs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.759381 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.759566 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-public-tls-certs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.762893 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.766626 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-config-data\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.770676 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l64c7\" (UniqueName: \"kubernetes.io/projected/5ee0c839-adff-484f-b5f5-1a836ead634f-kube-api-access-l64c7\") pod \"nova-api-0\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.850040 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.857527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-config-data\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.857664 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-scripts\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.857713 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.857735 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrtmf\" (UniqueName: \"kubernetes.io/projected/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-kube-api-access-wrtmf\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.862567 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-config-data\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.863361 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.869478 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-scripts\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:38 crc kubenswrapper[4894]: I1210 19:19:38.874545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrtmf\" (UniqueName: \"kubernetes.io/projected/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-kube-api-access-wrtmf\") pod \"nova-cell1-cell-mapping-9pwpt\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:39 crc kubenswrapper[4894]: I1210 19:19:39.163484 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:39 crc kubenswrapper[4894]: I1210 19:19:39.352095 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerStarted","Data":"76b1ce648639745b983d33e709731b8ca77403e75d2db7e4738be3faaaec83ec"} Dec 10 19:19:39 crc kubenswrapper[4894]: I1210 19:19:39.418280 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:39 crc kubenswrapper[4894]: I1210 19:19:39.557629 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25135a0c-3452-4c04-ab19-2757b663b60d" path="/var/lib/kubelet/pods/25135a0c-3452-4c04-ab19-2757b663b60d/volumes" Dec 10 19:19:39 crc kubenswrapper[4894]: I1210 19:19:39.679610 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9pwpt"] Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.362163 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerStarted","Data":"c7198e2ba584d88fc0dfaba7e963633f16ae41c5f98749410bbba96aa7feed30"} Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.363741 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ee0c839-adff-484f-b5f5-1a836ead634f","Type":"ContainerStarted","Data":"9a4330d32f88799bf9f854dc4be68d69ae162c737aae0d081b3abb437cce06a6"} Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.363804 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ee0c839-adff-484f-b5f5-1a836ead634f","Type":"ContainerStarted","Data":"6278558fd19518d4672e68ce0ac07918def0644368094f4bc06856d7a3bb3013"} Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.363819 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ee0c839-adff-484f-b5f5-1a836ead634f","Type":"ContainerStarted","Data":"7f37390fbc7a7c7c017556a722caaf83e53638b92050ebfcd5eae8de21efdd22"} Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.364986 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9pwpt" event={"ID":"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4","Type":"ContainerStarted","Data":"8976b2ff473c23acd67f8a5a995c7cc92487d6fc62a56378d4562438b200f239"} Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.365027 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9pwpt" event={"ID":"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4","Type":"ContainerStarted","Data":"4146518551d0a82c91c519d32f31f17c7958d8a828357cc6605ac42e8e500691"} Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.411575 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.411552938 podStartE2EDuration="2.411552938s" podCreationTimestamp="2025-12-10 19:19:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:40.394591949 +0000 UTC m=+1463.189439751" watchObservedRunningTime="2025-12-10 19:19:40.411552938 +0000 UTC m=+1463.206400740" Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.434121 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9pwpt" podStartSLOduration=2.434102969 podStartE2EDuration="2.434102969s" podCreationTimestamp="2025-12-10 19:19:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:40.426681337 +0000 UTC m=+1463.221529129" watchObservedRunningTime="2025-12-10 19:19:40.434102969 +0000 UTC m=+1463.228950761" Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.723002 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.789361 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c9cb78d75-nqzt2"] Dec 10 19:19:40 crc kubenswrapper[4894]: I1210 19:19:40.789641 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" containerName="dnsmasq-dns" containerID="cri-o://cc469c503b32241dd808394a83a6e9ef61915a403dc1a2be8f9537d87b7327d7" gracePeriod=10 Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.376582 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca6401e5-caa0-49b6-a281-b74b74190143" containerID="cc469c503b32241dd808394a83a6e9ef61915a403dc1a2be8f9537d87b7327d7" exitCode=0 Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.376687 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" event={"ID":"ca6401e5-caa0-49b6-a281-b74b74190143","Type":"ContainerDied","Data":"cc469c503b32241dd808394a83a6e9ef61915a403dc1a2be8f9537d87b7327d7"} Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.608260 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.649860 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-nb\") pod \"ca6401e5-caa0-49b6-a281-b74b74190143\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.649990 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-config\") pod \"ca6401e5-caa0-49b6-a281-b74b74190143\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.650035 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-svc\") pod \"ca6401e5-caa0-49b6-a281-b74b74190143\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.650116 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-swift-storage-0\") pod \"ca6401e5-caa0-49b6-a281-b74b74190143\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.650158 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k89kz\" (UniqueName: \"kubernetes.io/projected/ca6401e5-caa0-49b6-a281-b74b74190143-kube-api-access-k89kz\") pod \"ca6401e5-caa0-49b6-a281-b74b74190143\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.650359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-sb\") pod \"ca6401e5-caa0-49b6-a281-b74b74190143\" (UID: \"ca6401e5-caa0-49b6-a281-b74b74190143\") " Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.659206 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca6401e5-caa0-49b6-a281-b74b74190143-kube-api-access-k89kz" (OuterVolumeSpecName: "kube-api-access-k89kz") pod "ca6401e5-caa0-49b6-a281-b74b74190143" (UID: "ca6401e5-caa0-49b6-a281-b74b74190143"). InnerVolumeSpecName "kube-api-access-k89kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.754167 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k89kz\" (UniqueName: \"kubernetes.io/projected/ca6401e5-caa0-49b6-a281-b74b74190143-kube-api-access-k89kz\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.846238 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ca6401e5-caa0-49b6-a281-b74b74190143" (UID: "ca6401e5-caa0-49b6-a281-b74b74190143"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.860156 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.862243 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ca6401e5-caa0-49b6-a281-b74b74190143" (UID: "ca6401e5-caa0-49b6-a281-b74b74190143"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.869362 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ca6401e5-caa0-49b6-a281-b74b74190143" (UID: "ca6401e5-caa0-49b6-a281-b74b74190143"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.874287 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-config" (OuterVolumeSpecName: "config") pod "ca6401e5-caa0-49b6-a281-b74b74190143" (UID: "ca6401e5-caa0-49b6-a281-b74b74190143"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.888403 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ca6401e5-caa0-49b6-a281-b74b74190143" (UID: "ca6401e5-caa0-49b6-a281-b74b74190143"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.962069 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.962102 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.962114 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:41 crc kubenswrapper[4894]: I1210 19:19:41.962122 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca6401e5-caa0-49b6-a281-b74b74190143-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.391481 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerStarted","Data":"3b03f7d9609967ad8429ad9467681e2371c3154d6b19d88f8ae5d9240c0466d2"} Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.393104 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.395690 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" event={"ID":"ca6401e5-caa0-49b6-a281-b74b74190143","Type":"ContainerDied","Data":"62e84fc7c26de3760876f5c6883205ac0e76016ea76ae6f63a44f1c3097a51ef"} Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.395758 4894 scope.go:117] "RemoveContainer" containerID="cc469c503b32241dd808394a83a6e9ef61915a403dc1a2be8f9537d87b7327d7" Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.395961 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cb78d75-nqzt2" Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.431780 4894 scope.go:117] "RemoveContainer" containerID="9710d519a07e8e8982454ed02387beb1dde89fcba3bab3ab6c485542a59469ea" Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.448343 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.434042036 podStartE2EDuration="6.448301303s" podCreationTimestamp="2025-12-10 19:19:36 +0000 UTC" firstStartedPulling="2025-12-10 19:19:37.400184685 +0000 UTC m=+1460.195032477" lastFinishedPulling="2025-12-10 19:19:41.414443952 +0000 UTC m=+1464.209291744" observedRunningTime="2025-12-10 19:19:42.419033061 +0000 UTC m=+1465.213880863" watchObservedRunningTime="2025-12-10 19:19:42.448301303 +0000 UTC m=+1465.243149115" Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.473492 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c9cb78d75-nqzt2"] Dec 10 19:19:42 crc kubenswrapper[4894]: I1210 19:19:42.484730 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c9cb78d75-nqzt2"] Dec 10 19:19:43 crc kubenswrapper[4894]: I1210 19:19:43.554753 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" path="/var/lib/kubelet/pods/ca6401e5-caa0-49b6-a281-b74b74190143/volumes" Dec 10 19:19:45 crc kubenswrapper[4894]: I1210 19:19:45.442190 4894 generic.go:334] "Generic (PLEG): container finished" podID="2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" containerID="8976b2ff473c23acd67f8a5a995c7cc92487d6fc62a56378d4562438b200f239" exitCode=0 Dec 10 19:19:45 crc kubenswrapper[4894]: I1210 19:19:45.444174 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9pwpt" event={"ID":"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4","Type":"ContainerDied","Data":"8976b2ff473c23acd67f8a5a995c7cc92487d6fc62a56378d4562438b200f239"} Dec 10 19:19:46 crc kubenswrapper[4894]: I1210 19:19:46.963025 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.073284 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-config-data\") pod \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.073683 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-scripts\") pod \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.073978 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrtmf\" (UniqueName: \"kubernetes.io/projected/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-kube-api-access-wrtmf\") pod \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.074279 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-combined-ca-bundle\") pod \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\" (UID: \"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4\") " Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.080677 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-kube-api-access-wrtmf" (OuterVolumeSpecName: "kube-api-access-wrtmf") pod "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" (UID: "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4"). InnerVolumeSpecName "kube-api-access-wrtmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.081060 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-scripts" (OuterVolumeSpecName: "scripts") pod "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" (UID: "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.124402 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" (UID: "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.124752 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-config-data" (OuterVolumeSpecName: "config-data") pod "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" (UID: "2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.177209 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.177263 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.177286 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrtmf\" (UniqueName: \"kubernetes.io/projected/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-kube-api-access-wrtmf\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.177306 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.467929 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9pwpt" event={"ID":"2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4","Type":"ContainerDied","Data":"4146518551d0a82c91c519d32f31f17c7958d8a828357cc6605ac42e8e500691"} Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.467973 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4146518551d0a82c91c519d32f31f17c7958d8a828357cc6605ac42e8e500691" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.468005 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9pwpt" Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.672536 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.673122 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="19974223-7dae-46bb-a05d-378d0afddec8" containerName="nova-scheduler-scheduler" containerID="cri-o://97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" gracePeriod=30 Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.687446 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.687701 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-log" containerID="cri-o://6278558fd19518d4672e68ce0ac07918def0644368094f4bc06856d7a3bb3013" gracePeriod=30 Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.687815 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-api" containerID="cri-o://9a4330d32f88799bf9f854dc4be68d69ae162c737aae0d081b3abb437cce06a6" gracePeriod=30 Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.763325 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.763556 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-log" containerID="cri-o://bfe3e77139e89b0df51cc01a877d313db3a25da24ac58b928ea707ce32c7546c" gracePeriod=30 Dec 10 19:19:47 crc kubenswrapper[4894]: I1210 19:19:47.763666 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-metadata" containerID="cri-o://9efc7f1829d23141066e2b433f702a084a92952d196fc63d2553343ba969e812" gracePeriod=30 Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.483544 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerID="9a4330d32f88799bf9f854dc4be68d69ae162c737aae0d081b3abb437cce06a6" exitCode=0 Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.483584 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerID="6278558fd19518d4672e68ce0ac07918def0644368094f4bc06856d7a3bb3013" exitCode=143 Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.483609 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ee0c839-adff-484f-b5f5-1a836ead634f","Type":"ContainerDied","Data":"9a4330d32f88799bf9f854dc4be68d69ae162c737aae0d081b3abb437cce06a6"} Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.483649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ee0c839-adff-484f-b5f5-1a836ead634f","Type":"ContainerDied","Data":"6278558fd19518d4672e68ce0ac07918def0644368094f4bc06856d7a3bb3013"} Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.486338 4894 generic.go:334] "Generic (PLEG): container finished" podID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerID="bfe3e77139e89b0df51cc01a877d313db3a25da24ac58b928ea707ce32c7546c" exitCode=143 Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.486402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f498f90b-987c-4be3-9702-5ba84af2e03b","Type":"ContainerDied","Data":"bfe3e77139e89b0df51cc01a877d313db3a25da24ac58b928ea707ce32c7546c"} Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.771683 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.818597 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-config-data\") pod \"5ee0c839-adff-484f-b5f5-1a836ead634f\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.818714 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee0c839-adff-484f-b5f5-1a836ead634f-logs\") pod \"5ee0c839-adff-484f-b5f5-1a836ead634f\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.818826 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-public-tls-certs\") pod \"5ee0c839-adff-484f-b5f5-1a836ead634f\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.818902 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-internal-tls-certs\") pod \"5ee0c839-adff-484f-b5f5-1a836ead634f\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.818992 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-combined-ca-bundle\") pod \"5ee0c839-adff-484f-b5f5-1a836ead634f\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.819074 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l64c7\" (UniqueName: \"kubernetes.io/projected/5ee0c839-adff-484f-b5f5-1a836ead634f-kube-api-access-l64c7\") pod \"5ee0c839-adff-484f-b5f5-1a836ead634f\" (UID: \"5ee0c839-adff-484f-b5f5-1a836ead634f\") " Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.819138 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee0c839-adff-484f-b5f5-1a836ead634f-logs" (OuterVolumeSpecName: "logs") pod "5ee0c839-adff-484f-b5f5-1a836ead634f" (UID: "5ee0c839-adff-484f-b5f5-1a836ead634f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.819689 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee0c839-adff-484f-b5f5-1a836ead634f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.833948 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee0c839-adff-484f-b5f5-1a836ead634f-kube-api-access-l64c7" (OuterVolumeSpecName: "kube-api-access-l64c7") pod "5ee0c839-adff-484f-b5f5-1a836ead634f" (UID: "5ee0c839-adff-484f-b5f5-1a836ead634f"). InnerVolumeSpecName "kube-api-access-l64c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.867961 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ee0c839-adff-484f-b5f5-1a836ead634f" (UID: "5ee0c839-adff-484f-b5f5-1a836ead634f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.881135 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-config-data" (OuterVolumeSpecName: "config-data") pod "5ee0c839-adff-484f-b5f5-1a836ead634f" (UID: "5ee0c839-adff-484f-b5f5-1a836ead634f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.910012 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5ee0c839-adff-484f-b5f5-1a836ead634f" (UID: "5ee0c839-adff-484f-b5f5-1a836ead634f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.911859 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5ee0c839-adff-484f-b5f5-1a836ead634f" (UID: "5ee0c839-adff-484f-b5f5-1a836ead634f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.921209 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.921240 4894 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.921250 4894 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.921261 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee0c839-adff-484f-b5f5-1a836ead634f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:48 crc kubenswrapper[4894]: I1210 19:19:48.921273 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l64c7\" (UniqueName: \"kubernetes.io/projected/5ee0c839-adff-484f-b5f5-1a836ead634f-kube-api-access-l64c7\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.498269 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ee0c839-adff-484f-b5f5-1a836ead634f","Type":"ContainerDied","Data":"7f37390fbc7a7c7c017556a722caaf83e53638b92050ebfcd5eae8de21efdd22"} Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.498625 4894 scope.go:117] "RemoveContainer" containerID="9a4330d32f88799bf9f854dc4be68d69ae162c737aae0d081b3abb437cce06a6" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.498343 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.535808 4894 scope.go:117] "RemoveContainer" containerID="6278558fd19518d4672e68ce0ac07918def0644368094f4bc06856d7a3bb3013" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.536058 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.581381 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.581478 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:49 crc kubenswrapper[4894]: E1210 19:19:49.582820 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" containerName="nova-manage" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.582892 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" containerName="nova-manage" Dec 10 19:19:49 crc kubenswrapper[4894]: E1210 19:19:49.582932 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" containerName="dnsmasq-dns" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583002 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" containerName="dnsmasq-dns" Dec 10 19:19:49 crc kubenswrapper[4894]: E1210 19:19:49.583023 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-log" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583072 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-log" Dec 10 19:19:49 crc kubenswrapper[4894]: E1210 19:19:49.583094 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-api" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583107 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-api" Dec 10 19:19:49 crc kubenswrapper[4894]: E1210 19:19:49.583169 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" containerName="init" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583185 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" containerName="init" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583830 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-log" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583926 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca6401e5-caa0-49b6-a281-b74b74190143" containerName="dnsmasq-dns" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.583993 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" containerName="nova-api-api" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.584027 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" containerName="nova-manage" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.589127 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.592833 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.593071 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.593481 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.595146 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.642686 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-config-data\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.642773 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10fac35-af08-4efc-b7ec-e7a39bdf0245-logs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.642826 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.642954 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.643082 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvhvc\" (UniqueName: \"kubernetes.io/projected/c10fac35-af08-4efc-b7ec-e7a39bdf0245-kube-api-access-kvhvc\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.643141 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-public-tls-certs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.744412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.744499 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.744590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvhvc\" (UniqueName: \"kubernetes.io/projected/c10fac35-af08-4efc-b7ec-e7a39bdf0245-kube-api-access-kvhvc\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.744644 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-public-tls-certs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.744685 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-config-data\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.744726 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10fac35-af08-4efc-b7ec-e7a39bdf0245-logs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.745615 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10fac35-af08-4efc-b7ec-e7a39bdf0245-logs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.748780 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-public-tls-certs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.748812 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.748817 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-config-data\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.750569 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10fac35-af08-4efc-b7ec-e7a39bdf0245-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.763607 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvhvc\" (UniqueName: \"kubernetes.io/projected/c10fac35-af08-4efc-b7ec-e7a39bdf0245-kube-api-access-kvhvc\") pod \"nova-api-0\" (UID: \"c10fac35-af08-4efc-b7ec-e7a39bdf0245\") " pod="openstack/nova-api-0" Dec 10 19:19:49 crc kubenswrapper[4894]: I1210 19:19:49.937799 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:19:50 crc kubenswrapper[4894]: I1210 19:19:50.488217 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:19:50 crc kubenswrapper[4894]: W1210 19:19:50.493015 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc10fac35_af08_4efc_b7ec_e7a39bdf0245.slice/crio-44f7dacf477eccb42256b71b2a6f4bac8f637daf60f321565fd1f1126c75316e WatchSource:0}: Error finding container 44f7dacf477eccb42256b71b2a6f4bac8f637daf60f321565fd1f1126c75316e: Status 404 returned error can't find the container with id 44f7dacf477eccb42256b71b2a6f4bac8f637daf60f321565fd1f1126c75316e Dec 10 19:19:50 crc kubenswrapper[4894]: I1210 19:19:50.540168 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c10fac35-af08-4efc-b7ec-e7a39bdf0245","Type":"ContainerStarted","Data":"44f7dacf477eccb42256b71b2a6f4bac8f637daf60f321565fd1f1126c75316e"} Dec 10 19:19:51 crc kubenswrapper[4894]: E1210 19:19:51.163458 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4 is running failed: container process not found" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:19:51 crc kubenswrapper[4894]: E1210 19:19:51.164971 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4 is running failed: container process not found" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:19:51 crc kubenswrapper[4894]: E1210 19:19:51.165424 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4 is running failed: container process not found" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:19:51 crc kubenswrapper[4894]: E1210 19:19:51.165453 4894 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="19974223-7dae-46bb-a05d-378d0afddec8" containerName="nova-scheduler-scheduler" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.227243 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.292743 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-config-data\") pod \"19974223-7dae-46bb-a05d-378d0afddec8\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.292835 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgwjm\" (UniqueName: \"kubernetes.io/projected/19974223-7dae-46bb-a05d-378d0afddec8-kube-api-access-pgwjm\") pod \"19974223-7dae-46bb-a05d-378d0afddec8\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.292893 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-combined-ca-bundle\") pod \"19974223-7dae-46bb-a05d-378d0afddec8\" (UID: \"19974223-7dae-46bb-a05d-378d0afddec8\") " Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.297488 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19974223-7dae-46bb-a05d-378d0afddec8-kube-api-access-pgwjm" (OuterVolumeSpecName: "kube-api-access-pgwjm") pod "19974223-7dae-46bb-a05d-378d0afddec8" (UID: "19974223-7dae-46bb-a05d-378d0afddec8"). InnerVolumeSpecName "kube-api-access-pgwjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.324020 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-config-data" (OuterVolumeSpecName: "config-data") pod "19974223-7dae-46bb-a05d-378d0afddec8" (UID: "19974223-7dae-46bb-a05d-378d0afddec8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.337640 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19974223-7dae-46bb-a05d-378d0afddec8" (UID: "19974223-7dae-46bb-a05d-378d0afddec8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.398397 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.398663 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgwjm\" (UniqueName: \"kubernetes.io/projected/19974223-7dae-46bb-a05d-378d0afddec8-kube-api-access-pgwjm\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.398674 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19974223-7dae-46bb-a05d-378d0afddec8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.412456 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.215:8775/\": dial tcp 10.217.0.215:8775: connect: connection refused" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.412607 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.215:8775/\": dial tcp 10.217.0.215:8775: connect: connection refused" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.567607 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee0c839-adff-484f-b5f5-1a836ead634f" path="/var/lib/kubelet/pods/5ee0c839-adff-484f-b5f5-1a836ead634f/volumes" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.568542 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c10fac35-af08-4efc-b7ec-e7a39bdf0245","Type":"ContainerStarted","Data":"5f5919d048921f0bc8406b6aa10c753842dded4c3d4b2a7628671b2029bed800"} Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.568572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c10fac35-af08-4efc-b7ec-e7a39bdf0245","Type":"ContainerStarted","Data":"095d2281fc2dbbf76dfbc174ad779172c6fefc5cc29209fe2cbe5b3d6a06d05b"} Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.574549 4894 generic.go:334] "Generic (PLEG): container finished" podID="19974223-7dae-46bb-a05d-378d0afddec8" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" exitCode=0 Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.574639 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19974223-7dae-46bb-a05d-378d0afddec8","Type":"ContainerDied","Data":"97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4"} Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.574694 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19974223-7dae-46bb-a05d-378d0afddec8","Type":"ContainerDied","Data":"c6130adc10c3f0a4ae2495b0de4707339bc56c2d485a1bd21d4559f9683515bc"} Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.574713 4894 scope.go:117] "RemoveContainer" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.574831 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.615331 4894 generic.go:334] "Generic (PLEG): container finished" podID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerID="9efc7f1829d23141066e2b433f702a084a92952d196fc63d2553343ba969e812" exitCode=0 Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.615376 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f498f90b-987c-4be3-9702-5ba84af2e03b","Type":"ContainerDied","Data":"9efc7f1829d23141066e2b433f702a084a92952d196fc63d2553343ba969e812"} Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.643013 4894 scope.go:117] "RemoveContainer" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" Dec 10 19:19:51 crc kubenswrapper[4894]: E1210 19:19:51.645148 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4\": container with ID starting with 97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4 not found: ID does not exist" containerID="97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.645176 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4"} err="failed to get container status \"97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4\": rpc error: code = NotFound desc = could not find container \"97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4\": container with ID starting with 97b62c7388b9df75cf4c8af9b3e17a12cb88a1cdf9d6de801b78898d097f77b4 not found: ID does not exist" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.647519 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.647494514 podStartE2EDuration="2.647494514s" podCreationTimestamp="2025-12-10 19:19:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:51.62814964 +0000 UTC m=+1474.422997442" watchObservedRunningTime="2025-12-10 19:19:51.647494514 +0000 UTC m=+1474.442342306" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.704029 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.727318 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.746399 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:51 crc kubenswrapper[4894]: E1210 19:19:51.746914 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19974223-7dae-46bb-a05d-378d0afddec8" containerName="nova-scheduler-scheduler" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.746936 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="19974223-7dae-46bb-a05d-378d0afddec8" containerName="nova-scheduler-scheduler" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.747195 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="19974223-7dae-46bb-a05d-378d0afddec8" containerName="nova-scheduler-scheduler" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.748903 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.752710 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.755578 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.806689 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a21758b8-7fad-4aff-ab6e-131b65d205ab-config-data\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.806814 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a21758b8-7fad-4aff-ab6e-131b65d205ab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.806933 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4zkn\" (UniqueName: \"kubernetes.io/projected/a21758b8-7fad-4aff-ab6e-131b65d205ab-kube-api-access-m4zkn\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.908395 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a21758b8-7fad-4aff-ab6e-131b65d205ab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.908591 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4zkn\" (UniqueName: \"kubernetes.io/projected/a21758b8-7fad-4aff-ab6e-131b65d205ab-kube-api-access-m4zkn\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.908698 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a21758b8-7fad-4aff-ab6e-131b65d205ab-config-data\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.913317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a21758b8-7fad-4aff-ab6e-131b65d205ab-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.913445 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a21758b8-7fad-4aff-ab6e-131b65d205ab-config-data\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:51 crc kubenswrapper[4894]: I1210 19:19:51.928186 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4zkn\" (UniqueName: \"kubernetes.io/projected/a21758b8-7fad-4aff-ab6e-131b65d205ab-kube-api-access-m4zkn\") pod \"nova-scheduler-0\" (UID: \"a21758b8-7fad-4aff-ab6e-131b65d205ab\") " pod="openstack/nova-scheduler-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.065950 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.385975 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.421503 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f498f90b-987c-4be3-9702-5ba84af2e03b-logs\") pod \"f498f90b-987c-4be3-9702-5ba84af2e03b\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.421616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9m2j\" (UniqueName: \"kubernetes.io/projected/f498f90b-987c-4be3-9702-5ba84af2e03b-kube-api-access-m9m2j\") pod \"f498f90b-987c-4be3-9702-5ba84af2e03b\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.421665 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-nova-metadata-tls-certs\") pod \"f498f90b-987c-4be3-9702-5ba84af2e03b\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.421728 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-config-data\") pod \"f498f90b-987c-4be3-9702-5ba84af2e03b\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.421861 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-combined-ca-bundle\") pod \"f498f90b-987c-4be3-9702-5ba84af2e03b\" (UID: \"f498f90b-987c-4be3-9702-5ba84af2e03b\") " Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.422244 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f498f90b-987c-4be3-9702-5ba84af2e03b-logs" (OuterVolumeSpecName: "logs") pod "f498f90b-987c-4be3-9702-5ba84af2e03b" (UID: "f498f90b-987c-4be3-9702-5ba84af2e03b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.422467 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f498f90b-987c-4be3-9702-5ba84af2e03b-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.427610 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f498f90b-987c-4be3-9702-5ba84af2e03b-kube-api-access-m9m2j" (OuterVolumeSpecName: "kube-api-access-m9m2j") pod "f498f90b-987c-4be3-9702-5ba84af2e03b" (UID: "f498f90b-987c-4be3-9702-5ba84af2e03b"). InnerVolumeSpecName "kube-api-access-m9m2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.481748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-config-data" (OuterVolumeSpecName: "config-data") pod "f498f90b-987c-4be3-9702-5ba84af2e03b" (UID: "f498f90b-987c-4be3-9702-5ba84af2e03b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.481836 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f498f90b-987c-4be3-9702-5ba84af2e03b" (UID: "f498f90b-987c-4be3-9702-5ba84af2e03b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.510204 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.523539 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f498f90b-987c-4be3-9702-5ba84af2e03b" (UID: "f498f90b-987c-4be3-9702-5ba84af2e03b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.525183 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.525255 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9m2j\" (UniqueName: \"kubernetes.io/projected/f498f90b-987c-4be3-9702-5ba84af2e03b-kube-api-access-m9m2j\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.525270 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.525286 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f498f90b-987c-4be3-9702-5ba84af2e03b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.629766 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a21758b8-7fad-4aff-ab6e-131b65d205ab","Type":"ContainerStarted","Data":"842aceab815733a7b85121d0882bb89c5fc6b783836030b891b5b0ea97df0d77"} Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.633807 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.634112 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f498f90b-987c-4be3-9702-5ba84af2e03b","Type":"ContainerDied","Data":"71589a9c6075e7779726a11782413df6abdd2fad2baf63762cb20dd71479bede"} Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.634183 4894 scope.go:117] "RemoveContainer" containerID="9efc7f1829d23141066e2b433f702a084a92952d196fc63d2553343ba969e812" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.683640 4894 scope.go:117] "RemoveContainer" containerID="bfe3e77139e89b0df51cc01a877d313db3a25da24ac58b928ea707ce32c7546c" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.683752 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.700290 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.723066 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:52 crc kubenswrapper[4894]: E1210 19:19:52.723572 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-metadata" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.723600 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-metadata" Dec 10 19:19:52 crc kubenswrapper[4894]: E1210 19:19:52.723638 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-log" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.723648 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-log" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.723920 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-metadata" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.723949 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" containerName="nova-metadata-log" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.725110 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.728810 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.729568 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.767273 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.832336 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.832395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h89dr\" (UniqueName: \"kubernetes.io/projected/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-kube-api-access-h89dr\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.832424 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-config-data\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.832479 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-logs\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.832514 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.934310 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h89dr\" (UniqueName: \"kubernetes.io/projected/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-kube-api-access-h89dr\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.934384 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-config-data\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.934473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-logs\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.934524 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.934685 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.935580 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-logs\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.938580 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.940518 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-config-data\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.941027 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:52 crc kubenswrapper[4894]: I1210 19:19:52.950373 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h89dr\" (UniqueName: \"kubernetes.io/projected/e4c4f31f-885b-4386-b6ae-2a0eb625b8d4-kube-api-access-h89dr\") pod \"nova-metadata-0\" (UID: \"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4\") " pod="openstack/nova-metadata-0" Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.076349 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.564187 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19974223-7dae-46bb-a05d-378d0afddec8" path="/var/lib/kubelet/pods/19974223-7dae-46bb-a05d-378d0afddec8/volumes" Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.566051 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f498f90b-987c-4be3-9702-5ba84af2e03b" path="/var/lib/kubelet/pods/f498f90b-987c-4be3-9702-5ba84af2e03b/volumes" Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.566814 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.647433 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4","Type":"ContainerStarted","Data":"6e4667f2baa9c038690416287c8ac27f85e4635b8c6f28252a24f759c1440124"} Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.651059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a21758b8-7fad-4aff-ab6e-131b65d205ab","Type":"ContainerStarted","Data":"2ab27f3abdeb806b4ce73cf41debd9da0be3ee265189c6949ec26e0316511152"} Dec 10 19:19:53 crc kubenswrapper[4894]: I1210 19:19:53.685805 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.685784391 podStartE2EDuration="2.685784391s" podCreationTimestamp="2025-12-10 19:19:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:53.672632865 +0000 UTC m=+1476.467480667" watchObservedRunningTime="2025-12-10 19:19:53.685784391 +0000 UTC m=+1476.480632203" Dec 10 19:19:54 crc kubenswrapper[4894]: I1210 19:19:54.675264 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4","Type":"ContainerStarted","Data":"3427e85ad7dd6248267dad7d90d5dc50fbc45a3c7a9916f02b2f55d3f62d634a"} Dec 10 19:19:54 crc kubenswrapper[4894]: I1210 19:19:54.675698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e4c4f31f-885b-4386-b6ae-2a0eb625b8d4","Type":"ContainerStarted","Data":"b8fee4e0735e6170bf0dd52e5f090f6ae5760a80dafb0fe15010508d0f376229"} Dec 10 19:19:54 crc kubenswrapper[4894]: I1210 19:19:54.699735 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.699713143 podStartE2EDuration="2.699713143s" podCreationTimestamp="2025-12-10 19:19:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:54.699637451 +0000 UTC m=+1477.494485243" watchObservedRunningTime="2025-12-10 19:19:54.699713143 +0000 UTC m=+1477.494560965" Dec 10 19:19:57 crc kubenswrapper[4894]: I1210 19:19:57.067010 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 19:19:58 crc kubenswrapper[4894]: I1210 19:19:58.077800 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:19:58 crc kubenswrapper[4894]: I1210 19:19:58.078084 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:19:59 crc kubenswrapper[4894]: I1210 19:19:59.938609 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:19:59 crc kubenswrapper[4894]: I1210 19:19:59.938967 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:20:00 crc kubenswrapper[4894]: I1210 19:20:00.958096 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c10fac35-af08-4efc-b7ec-e7a39bdf0245" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:00 crc kubenswrapper[4894]: I1210 19:20:00.958298 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c10fac35-af08-4efc-b7ec-e7a39bdf0245" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.225:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:02 crc kubenswrapper[4894]: I1210 19:20:02.067089 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 19:20:02 crc kubenswrapper[4894]: I1210 19:20:02.096181 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 19:20:02 crc kubenswrapper[4894]: I1210 19:20:02.821246 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 19:20:03 crc kubenswrapper[4894]: I1210 19:20:03.077373 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:20:03 crc kubenswrapper[4894]: I1210 19:20:03.077422 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:20:04 crc kubenswrapper[4894]: I1210 19:20:04.089013 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e4c4f31f-885b-4386-b6ae-2a0eb625b8d4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.227:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:04 crc kubenswrapper[4894]: I1210 19:20:04.089011 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e4c4f31f-885b-4386-b6ae-2a0eb625b8d4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.227:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:06 crc kubenswrapper[4894]: I1210 19:20:06.704448 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 19:20:09 crc kubenswrapper[4894]: I1210 19:20:09.948635 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:20:09 crc kubenswrapper[4894]: I1210 19:20:09.949124 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:20:09 crc kubenswrapper[4894]: I1210 19:20:09.949568 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:20:09 crc kubenswrapper[4894]: I1210 19:20:09.949623 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:20:09 crc kubenswrapper[4894]: I1210 19:20:09.954445 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:20:09 crc kubenswrapper[4894]: I1210 19:20:09.954704 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:20:13 crc kubenswrapper[4894]: I1210 19:20:13.083882 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:20:13 crc kubenswrapper[4894]: I1210 19:20:13.084195 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:20:13 crc kubenswrapper[4894]: I1210 19:20:13.089932 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:20:13 crc kubenswrapper[4894]: I1210 19:20:13.091717 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.007184 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-vprjr"] Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.016729 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-vprjr"] Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.112523 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-zc7gw"] Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.114328 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.118321 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.124521 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-zc7gw"] Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.287894 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-combined-ca-bundle\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.288218 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-certs\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.288252 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-scripts\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.288274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw29b\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-kube-api-access-nw29b\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.288413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.390177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.390284 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-combined-ca-bundle\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.390313 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-certs\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.390347 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-scripts\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.390374 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw29b\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-kube-api-access-nw29b\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.396870 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-certs\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.397346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-scripts\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.413645 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-combined-ca-bundle\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.414584 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.416342 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw29b\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-kube-api-access-nw29b\") pod \"cloudkitty-db-sync-zc7gw\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.498132 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.556923 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc3ed2c8-9e2c-443b-bc22-a39397986e13" path="/var/lib/kubelet/pods/bc3ed2c8-9e2c-443b-bc22-a39397986e13/volumes" Dec 10 19:20:23 crc kubenswrapper[4894]: I1210 19:20:23.959401 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-zc7gw"] Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.014508 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zc7gw" event={"ID":"f972529a-80b2-4d4c-a9ed-1616052d506e","Type":"ContainerStarted","Data":"0794c2e71b6dc9b9a790ccbd1426d44e8edac01b0a48a1a5139ee70b8eba5a0b"} Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.583661 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.584306 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-central-agent" containerID="cri-o://3729d19ec9376b06cac616f59a5089e344b9b3bd31c4c0a9d3637082a47b8f13" gracePeriod=30 Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.584454 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="proxy-httpd" containerID="cri-o://3b03f7d9609967ad8429ad9467681e2371c3154d6b19d88f8ae5d9240c0466d2" gracePeriod=30 Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.584507 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="sg-core" containerID="cri-o://c7198e2ba584d88fc0dfaba7e963633f16ae41c5f98749410bbba96aa7feed30" gracePeriod=30 Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.584550 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-notification-agent" containerID="cri-o://76b1ce648639745b983d33e709731b8ca77403e75d2db7e4738be3faaaec83ec" gracePeriod=30 Dec 10 19:20:24 crc kubenswrapper[4894]: I1210 19:20:24.843389 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.056601 4894 generic.go:334] "Generic (PLEG): container finished" podID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerID="3b03f7d9609967ad8429ad9467681e2371c3154d6b19d88f8ae5d9240c0466d2" exitCode=0 Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.056633 4894 generic.go:334] "Generic (PLEG): container finished" podID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerID="c7198e2ba584d88fc0dfaba7e963633f16ae41c5f98749410bbba96aa7feed30" exitCode=2 Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.056683 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerDied","Data":"3b03f7d9609967ad8429ad9467681e2371c3154d6b19d88f8ae5d9240c0466d2"} Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.056727 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerDied","Data":"c7198e2ba584d88fc0dfaba7e963633f16ae41c5f98749410bbba96aa7feed30"} Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.058065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zc7gw" event={"ID":"f972529a-80b2-4d4c-a9ed-1616052d506e","Type":"ContainerStarted","Data":"a699da97fb3212b780700797a9bdc9ff5a5855c7df869e685cfe7b2046b818a9"} Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.091748 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-zc7gw" podStartSLOduration=1.8910978040000002 podStartE2EDuration="2.091727215s" podCreationTimestamp="2025-12-10 19:20:23 +0000 UTC" firstStartedPulling="2025-12-10 19:20:23.963167982 +0000 UTC m=+1506.758015784" lastFinishedPulling="2025-12-10 19:20:24.163797403 +0000 UTC m=+1506.958645195" observedRunningTime="2025-12-10 19:20:25.089260546 +0000 UTC m=+1507.884108338" watchObservedRunningTime="2025-12-10 19:20:25.091727215 +0000 UTC m=+1507.886575007" Dec 10 19:20:25 crc kubenswrapper[4894]: I1210 19:20:25.735330 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:20:26 crc kubenswrapper[4894]: I1210 19:20:26.068993 4894 generic.go:334] "Generic (PLEG): container finished" podID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerID="3729d19ec9376b06cac616f59a5089e344b9b3bd31c4c0a9d3637082a47b8f13" exitCode=0 Dec 10 19:20:26 crc kubenswrapper[4894]: I1210 19:20:26.069064 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerDied","Data":"3729d19ec9376b06cac616f59a5089e344b9b3bd31c4c0a9d3637082a47b8f13"} Dec 10 19:20:27 crc kubenswrapper[4894]: I1210 19:20:27.079123 4894 generic.go:334] "Generic (PLEG): container finished" podID="f972529a-80b2-4d4c-a9ed-1616052d506e" containerID="a699da97fb3212b780700797a9bdc9ff5a5855c7df869e685cfe7b2046b818a9" exitCode=0 Dec 10 19:20:27 crc kubenswrapper[4894]: I1210 19:20:27.079205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zc7gw" event={"ID":"f972529a-80b2-4d4c-a9ed-1616052d506e","Type":"ContainerDied","Data":"a699da97fb3212b780700797a9bdc9ff5a5855c7df869e685cfe7b2046b818a9"} Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.092824 4894 generic.go:334] "Generic (PLEG): container finished" podID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerID="76b1ce648639745b983d33e709731b8ca77403e75d2db7e4738be3faaaec83ec" exitCode=0 Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.092884 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerDied","Data":"76b1ce648639745b983d33e709731b8ca77403e75d2db7e4738be3faaaec83ec"} Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.571005 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.700996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-log-httpd\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-config-data\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701198 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-ceilometer-tls-certs\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701241 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-run-httpd\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701295 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-sg-core-conf-yaml\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701330 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-combined-ca-bundle\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbpnj\" (UniqueName: \"kubernetes.io/projected/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-kube-api-access-cbpnj\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.701399 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-scripts\") pod \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\" (UID: \"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.702359 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.702721 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.709946 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-kube-api-access-cbpnj" (OuterVolumeSpecName: "kube-api-access-cbpnj") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "kube-api-access-cbpnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.722451 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-scripts" (OuterVolumeSpecName: "scripts") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.729300 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.754446 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.781787 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.803470 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.803520 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.803538 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.803554 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.803570 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbpnj\" (UniqueName: \"kubernetes.io/projected/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-kube-api-access-cbpnj\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.803585 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.828981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.857201 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-config-data" (OuterVolumeSpecName: "config-data") pod "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" (UID: "c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.904610 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-combined-ca-bundle\") pod \"f972529a-80b2-4d4c-a9ed-1616052d506e\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.904794 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data\") pod \"f972529a-80b2-4d4c-a9ed-1616052d506e\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.904865 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-scripts\") pod \"f972529a-80b2-4d4c-a9ed-1616052d506e\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.904894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw29b\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-kube-api-access-nw29b\") pod \"f972529a-80b2-4d4c-a9ed-1616052d506e\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.904971 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-certs\") pod \"f972529a-80b2-4d4c-a9ed-1616052d506e\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.905410 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.905427 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.923109 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-scripts" (OuterVolumeSpecName: "scripts") pod "f972529a-80b2-4d4c-a9ed-1616052d506e" (UID: "f972529a-80b2-4d4c-a9ed-1616052d506e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.925119 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-kube-api-access-nw29b" (OuterVolumeSpecName: "kube-api-access-nw29b") pod "f972529a-80b2-4d4c-a9ed-1616052d506e" (UID: "f972529a-80b2-4d4c-a9ed-1616052d506e"). InnerVolumeSpecName "kube-api-access-nw29b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.930337 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-certs" (OuterVolumeSpecName: "certs") pod "f972529a-80b2-4d4c-a9ed-1616052d506e" (UID: "f972529a-80b2-4d4c-a9ed-1616052d506e"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:28 crc kubenswrapper[4894]: E1210 19:20:28.938178 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data podName:f972529a-80b2-4d4c-a9ed-1616052d506e nodeName:}" failed. No retries permitted until 2025-12-10 19:20:29.43815178 +0000 UTC m=+1512.232999572 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data") pod "f972529a-80b2-4d4c-a9ed-1616052d506e" (UID: "f972529a-80b2-4d4c-a9ed-1616052d506e") : error deleting /var/lib/kubelet/pods/f972529a-80b2-4d4c-a9ed-1616052d506e/volume-subpaths: remove /var/lib/kubelet/pods/f972529a-80b2-4d4c-a9ed-1616052d506e/volume-subpaths: no such file or directory Dec 10 19:20:28 crc kubenswrapper[4894]: I1210 19:20:28.941708 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f972529a-80b2-4d4c-a9ed-1616052d506e" (UID: "f972529a-80b2-4d4c-a9ed-1616052d506e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.010436 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw29b\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-kube-api-access-nw29b\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.010477 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f972529a-80b2-4d4c-a9ed-1616052d506e-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.010491 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.010501 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.101780 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zc7gw" event={"ID":"f972529a-80b2-4d4c-a9ed-1616052d506e","Type":"ContainerDied","Data":"0794c2e71b6dc9b9a790ccbd1426d44e8edac01b0a48a1a5139ee70b8eba5a0b"} Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.101818 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0794c2e71b6dc9b9a790ccbd1426d44e8edac01b0a48a1a5139ee70b8eba5a0b" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.101861 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zc7gw" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.105016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e","Type":"ContainerDied","Data":"a7b2b495f8d429b02bce7ee638dff34c6fbf9082b44e7c64f465670747bd1c1e"} Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.105077 4894 scope.go:117] "RemoveContainer" containerID="3b03f7d9609967ad8429ad9467681e2371c3154d6b19d88f8ae5d9240c0466d2" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.105093 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.131618 4894 scope.go:117] "RemoveContainer" containerID="c7198e2ba584d88fc0dfaba7e963633f16ae41c5f98749410bbba96aa7feed30" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.154915 4894 scope.go:117] "RemoveContainer" containerID="76b1ce648639745b983d33e709731b8ca77403e75d2db7e4738be3faaaec83ec" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.202555 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.203333 4894 scope.go:117] "RemoveContainer" containerID="3729d19ec9376b06cac616f59a5089e344b9b3bd31c4c0a9d3637082a47b8f13" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234064 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234127 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:29 crc kubenswrapper[4894]: E1210 19:20:29.234507 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-notification-agent" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234519 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-notification-agent" Dec 10 19:20:29 crc kubenswrapper[4894]: E1210 19:20:29.234532 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="proxy-httpd" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234539 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="proxy-httpd" Dec 10 19:20:29 crc kubenswrapper[4894]: E1210 19:20:29.234582 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="sg-core" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234589 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="sg-core" Dec 10 19:20:29 crc kubenswrapper[4894]: E1210 19:20:29.234604 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-central-agent" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234611 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-central-agent" Dec 10 19:20:29 crc kubenswrapper[4894]: E1210 19:20:29.234627 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f972529a-80b2-4d4c-a9ed-1616052d506e" containerName="cloudkitty-db-sync" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234634 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f972529a-80b2-4d4c-a9ed-1616052d506e" containerName="cloudkitty-db-sync" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234871 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-central-agent" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234889 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="sg-core" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234901 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f972529a-80b2-4d4c-a9ed-1616052d506e" containerName="cloudkitty-db-sync" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234916 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="ceilometer-notification-agent" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.234931 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" containerName="proxy-httpd" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.236785 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.236886 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259445 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt5qk\" (UniqueName: \"kubernetes.io/projected/10e3654c-3444-4fd9-9163-0b93a2089f15-kube-api-access-lt5qk\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259493 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10e3654c-3444-4fd9-9163-0b93a2089f15-log-httpd\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10e3654c-3444-4fd9-9163-0b93a2089f15-run-httpd\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259554 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259585 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-scripts\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259621 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-config-data\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259666 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.259740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.261292 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.261354 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.261538 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.339145 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-sprhg"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.348272 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-sprhg"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.356528 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-8r6r4"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.358131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361274 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-certs\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361354 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-config-data\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361373 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361396 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361433 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-scripts\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361463 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv64w\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-kube-api-access-xv64w\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361494 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt5qk\" (UniqueName: \"kubernetes.io/projected/10e3654c-3444-4fd9-9163-0b93a2089f15-kube-api-access-lt5qk\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361512 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10e3654c-3444-4fd9-9163-0b93a2089f15-log-httpd\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10e3654c-3444-4fd9-9163-0b93a2089f15-run-httpd\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361554 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361578 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-scripts\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.361604 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-config-data\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.363791 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10e3654c-3444-4fd9-9163-0b93a2089f15-log-httpd\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.368709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.369041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10e3654c-3444-4fd9-9163-0b93a2089f15-run-httpd\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.371309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-config-data\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.371364 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.371716 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8r6r4"] Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.376145 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-scripts\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.379217 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/10e3654c-3444-4fd9-9163-0b93a2089f15-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.392346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt5qk\" (UniqueName: \"kubernetes.io/projected/10e3654c-3444-4fd9-9163-0b93a2089f15-kube-api-access-lt5qk\") pod \"ceilometer-0\" (UID: \"10e3654c-3444-4fd9-9163-0b93a2089f15\") " pod="openstack/ceilometer-0" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.462706 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data\") pod \"f972529a-80b2-4d4c-a9ed-1616052d506e\" (UID: \"f972529a-80b2-4d4c-a9ed-1616052d506e\") " Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.463226 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-certs\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.463271 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-config-data\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.463288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.463336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-scripts\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.463366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv64w\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-kube-api-access-xv64w\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.470557 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-scripts\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.472320 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data" (OuterVolumeSpecName: "config-data") pod "f972529a-80b2-4d4c-a9ed-1616052d506e" (UID: "f972529a-80b2-4d4c-a9ed-1616052d506e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.474741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-certs\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.480965 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-config-data\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.489599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv64w\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-kube-api-access-xv64w\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.496616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-combined-ca-bundle\") pod \"cloudkitty-storageinit-8r6r4\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.527065 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="rabbitmq" containerID="cri-o://34c65a66b3db9b6eac6c97117490d462700c7f0cea9df74fa545c757a24f8484" gracePeriod=604796 Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.556580 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36157d9-6d04-40f0-a303-0ff6853ac239" path="/var/lib/kubelet/pods/b36157d9-6d04-40f0-a303-0ff6853ac239/volumes" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.557201 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e" path="/var/lib/kubelet/pods/c651b606-aadd-43d6-ba9c-a1c4bcd3fa7e/volumes" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.564853 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f972529a-80b2-4d4c-a9ed-1616052d506e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.595756 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:29 crc kubenswrapper[4894]: I1210 19:20:29.605494 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:30 crc kubenswrapper[4894]: I1210 19:20:30.102526 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:30 crc kubenswrapper[4894]: I1210 19:20:30.123189 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10e3654c-3444-4fd9-9163-0b93a2089f15","Type":"ContainerStarted","Data":"0af64979013dae71d03cfc57d81ddd8fbe5daf4d20436ed4014649b5f1202b9f"} Dec 10 19:20:30 crc kubenswrapper[4894]: I1210 19:20:30.196974 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8r6r4"] Dec 10 19:20:30 crc kubenswrapper[4894]: W1210 19:20:30.214762 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81df0908_03a5_439b_858d_e6e78ca7cce1.slice/crio-ea442c39249ac4a672b127cf3c5838c5c32d185c52811189fe65af1bde613328 WatchSource:0}: Error finding container ea442c39249ac4a672b127cf3c5838c5c32d185c52811189fe65af1bde613328: Status 404 returned error can't find the container with id ea442c39249ac4a672b127cf3c5838c5c32d185c52811189fe65af1bde613328 Dec 10 19:20:30 crc kubenswrapper[4894]: I1210 19:20:30.437481 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="rabbitmq" containerID="cri-o://7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c" gracePeriod=604796 Dec 10 19:20:31 crc kubenswrapper[4894]: I1210 19:20:31.134203 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r6r4" event={"ID":"81df0908-03a5-439b-858d-e6e78ca7cce1","Type":"ContainerStarted","Data":"ade95ffd9ee9fe93a57925519398a4c08dd6ac8215306564fad730aff70cf28e"} Dec 10 19:20:31 crc kubenswrapper[4894]: I1210 19:20:31.134534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r6r4" event={"ID":"81df0908-03a5-439b-858d-e6e78ca7cce1","Type":"ContainerStarted","Data":"ea442c39249ac4a672b127cf3c5838c5c32d185c52811189fe65af1bde613328"} Dec 10 19:20:31 crc kubenswrapper[4894]: I1210 19:20:31.150485 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-8r6r4" podStartSLOduration=2.150464507 podStartE2EDuration="2.150464507s" podCreationTimestamp="2025-12-10 19:20:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:31.147921027 +0000 UTC m=+1513.942768829" watchObservedRunningTime="2025-12-10 19:20:31.150464507 +0000 UTC m=+1513.945312299" Dec 10 19:20:33 crc kubenswrapper[4894]: I1210 19:20:33.182494 4894 generic.go:334] "Generic (PLEG): container finished" podID="81df0908-03a5-439b-858d-e6e78ca7cce1" containerID="ade95ffd9ee9fe93a57925519398a4c08dd6ac8215306564fad730aff70cf28e" exitCode=0 Dec 10 19:20:33 crc kubenswrapper[4894]: I1210 19:20:33.182583 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r6r4" event={"ID":"81df0908-03a5-439b-858d-e6e78ca7cce1","Type":"ContainerDied","Data":"ade95ffd9ee9fe93a57925519398a4c08dd6ac8215306564fad730aff70cf28e"} Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.194176 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10e3654c-3444-4fd9-9163-0b93a2089f15","Type":"ContainerStarted","Data":"25e2adb944030d10f17586b0d3dfb697c0a7240c2030809f0e102a6bf12f8661"} Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.730426 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.779947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv64w\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-kube-api-access-xv64w\") pod \"81df0908-03a5-439b-858d-e6e78ca7cce1\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.780192 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-config-data\") pod \"81df0908-03a5-439b-858d-e6e78ca7cce1\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.780389 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-scripts\") pod \"81df0908-03a5-439b-858d-e6e78ca7cce1\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.780487 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-combined-ca-bundle\") pod \"81df0908-03a5-439b-858d-e6e78ca7cce1\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.780758 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-certs\") pod \"81df0908-03a5-439b-858d-e6e78ca7cce1\" (UID: \"81df0908-03a5-439b-858d-e6e78ca7cce1\") " Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.786548 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-scripts" (OuterVolumeSpecName: "scripts") pod "81df0908-03a5-439b-858d-e6e78ca7cce1" (UID: "81df0908-03a5-439b-858d-e6e78ca7cce1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.787384 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-kube-api-access-xv64w" (OuterVolumeSpecName: "kube-api-access-xv64w") pod "81df0908-03a5-439b-858d-e6e78ca7cce1" (UID: "81df0908-03a5-439b-858d-e6e78ca7cce1"). InnerVolumeSpecName "kube-api-access-xv64w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.788025 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv64w\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-kube-api-access-xv64w\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.788097 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.803715 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-certs" (OuterVolumeSpecName: "certs") pod "81df0908-03a5-439b-858d-e6e78ca7cce1" (UID: "81df0908-03a5-439b-858d-e6e78ca7cce1"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.816501 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81df0908-03a5-439b-858d-e6e78ca7cce1" (UID: "81df0908-03a5-439b-858d-e6e78ca7cce1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.838464 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-config-data" (OuterVolumeSpecName: "config-data") pod "81df0908-03a5-439b-858d-e6e78ca7cce1" (UID: "81df0908-03a5-439b-858d-e6e78ca7cce1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.889821 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.889874 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/81df0908-03a5-439b-858d-e6e78ca7cce1-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4894]: I1210 19:20:34.889884 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81df0908-03a5-439b-858d-e6e78ca7cce1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.213425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8r6r4" event={"ID":"81df0908-03a5-439b-858d-e6e78ca7cce1","Type":"ContainerDied","Data":"ea442c39249ac4a672b127cf3c5838c5c32d185c52811189fe65af1bde613328"} Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.213898 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8r6r4" Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.214555 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea442c39249ac4a672b127cf3c5838c5c32d185c52811189fe65af1bde613328" Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.225015 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10e3654c-3444-4fd9-9163-0b93a2089f15","Type":"ContainerStarted","Data":"adb16e8548b83d0486519159cf0e19cb1dbbf72309a8f127fde28921d140b48f"} Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.338025 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.338279 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="228bcf93-2f3d-461a-a006-b5429740bf6d" containerName="cloudkitty-proc" containerID="cri-o://c2c50ad9b4115480c242cbd5f35d9418373a3cb54dbd904a4bf68dc7a14accad" gracePeriod=30 Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.350830 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.351129 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api-log" containerID="cri-o://885c37f3498f8bbf498badf4e3529e8cc0959aaa9cd8b6f56b38a0a27300bf4e" gracePeriod=30 Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.351193 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api" containerID="cri-o://fcd772eee6a7ebb5c88934d022ebd48ecddd199bbee9c9360156d698935d92e1" gracePeriod=30 Dec 10 19:20:35 crc kubenswrapper[4894]: I1210 19:20:35.824319 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.106:5671: connect: connection refused" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.106922 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.271235 4894 generic.go:334] "Generic (PLEG): container finished" podID="97355ddc-532e-46a2-a568-31f4a5e07030" containerID="34c65a66b3db9b6eac6c97117490d462700c7f0cea9df74fa545c757a24f8484" exitCode=0 Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.271320 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97355ddc-532e-46a2-a568-31f4a5e07030","Type":"ContainerDied","Data":"34c65a66b3db9b6eac6c97117490d462700c7f0cea9df74fa545c757a24f8484"} Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.271347 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97355ddc-532e-46a2-a568-31f4a5e07030","Type":"ContainerDied","Data":"d17bbf1ddf0f3043f97b69700cdf0d4b10ed61ce020314125ab026833e5a2c1f"} Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.271359 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d17bbf1ddf0f3043f97b69700cdf0d4b10ed61ce020314125ab026833e5a2c1f" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.305015 4894 generic.go:334] "Generic (PLEG): container finished" podID="228bcf93-2f3d-461a-a006-b5429740bf6d" containerID="c2c50ad9b4115480c242cbd5f35d9418373a3cb54dbd904a4bf68dc7a14accad" exitCode=0 Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.305083 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"228bcf93-2f3d-461a-a006-b5429740bf6d","Type":"ContainerDied","Data":"c2c50ad9b4115480c242cbd5f35d9418373a3cb54dbd904a4bf68dc7a14accad"} Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.342151 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.342968 4894 generic.go:334] "Generic (PLEG): container finished" podID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerID="885c37f3498f8bbf498badf4e3529e8cc0959aaa9cd8b6f56b38a0a27300bf4e" exitCode=143 Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.343022 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"38368ffb-25fe-478c-a1ab-5ca2b4a0c376","Type":"ContainerDied","Data":"885c37f3498f8bbf498badf4e3529e8cc0959aaa9cd8b6f56b38a0a27300bf4e"} Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.381015 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10e3654c-3444-4fd9-9163-0b93a2089f15","Type":"ContainerStarted","Data":"d48a06b2b8016d9472c2d933b92dcc97cda4e7daaed433a2295ea390c9df1cf5"} Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.429563 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kb2m\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-kube-api-access-2kb2m\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.429642 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-plugins-conf\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.429661 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97355ddc-532e-46a2-a568-31f4a5e07030-erlang-cookie-secret\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.429709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97355ddc-532e-46a2-a568-31f4a5e07030-pod-info\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.429742 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-tls\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.430455 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.430592 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-server-conf\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.430772 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-plugins\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.430871 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-config-data\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.430932 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-confd\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.430962 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-erlang-cookie\") pod \"97355ddc-532e-46a2-a568-31f4a5e07030\" (UID: \"97355ddc-532e-46a2-a568-31f4a5e07030\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.431717 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.436656 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.437121 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.448993 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-kube-api-access-2kb2m" (OuterVolumeSpecName: "kube-api-access-2kb2m") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "kube-api-access-2kb2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.461014 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.466036 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97355ddc-532e-46a2-a568-31f4a5e07030-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.471019 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/97355ddc-532e-46a2-a568-31f4a5e07030-pod-info" (OuterVolumeSpecName: "pod-info") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.502800 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-config-data" (OuterVolumeSpecName: "config-data") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534342 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534617 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kb2m\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-kube-api-access-2kb2m\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534628 4894 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534638 4894 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97355ddc-532e-46a2-a568-31f4a5e07030-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534646 4894 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97355ddc-532e-46a2-a568-31f4a5e07030-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534653 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534660 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.534668 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.539127 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b" (OuterVolumeSpecName: "persistence") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "pvc-d24fd592-c53c-42b1-8849-8a514156219b". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.573105 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-server-conf" (OuterVolumeSpecName: "server-conf") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.618236 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "97355ddc-532e-46a2-a568-31f4a5e07030" (UID: "97355ddc-532e-46a2-a568-31f4a5e07030"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.636349 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97355ddc-532e-46a2-a568-31f4a5e07030-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.636406 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") on node \"crc\" " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.636442 4894 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97355ddc-532e-46a2-a568-31f4a5e07030-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.679686 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.680960 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-d24fd592-c53c-42b1-8849-8a514156219b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b") on node "crc" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.738572 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.761241 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.941219 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-certs\") pod \"228bcf93-2f3d-461a-a006-b5429740bf6d\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.941523 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-combined-ca-bundle\") pod \"228bcf93-2f3d-461a-a006-b5429740bf6d\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.941585 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wc4l\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-kube-api-access-9wc4l\") pod \"228bcf93-2f3d-461a-a006-b5429740bf6d\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.941608 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-scripts\") pod \"228bcf93-2f3d-461a-a006-b5429740bf6d\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.941888 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data-custom\") pod \"228bcf93-2f3d-461a-a006-b5429740bf6d\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.941944 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data\") pod \"228bcf93-2f3d-461a-a006-b5429740bf6d\" (UID: \"228bcf93-2f3d-461a-a006-b5429740bf6d\") " Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.945155 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-scripts" (OuterVolumeSpecName: "scripts") pod "228bcf93-2f3d-461a-a006-b5429740bf6d" (UID: "228bcf93-2f3d-461a-a006-b5429740bf6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.946023 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-certs" (OuterVolumeSpecName: "certs") pod "228bcf93-2f3d-461a-a006-b5429740bf6d" (UID: "228bcf93-2f3d-461a-a006-b5429740bf6d"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.947067 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-kube-api-access-9wc4l" (OuterVolumeSpecName: "kube-api-access-9wc4l") pod "228bcf93-2f3d-461a-a006-b5429740bf6d" (UID: "228bcf93-2f3d-461a-a006-b5429740bf6d"). InnerVolumeSpecName "kube-api-access-9wc4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.981747 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "228bcf93-2f3d-461a-a006-b5429740bf6d" (UID: "228bcf93-2f3d-461a-a006-b5429740bf6d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4894]: I1210 19:20:36.998315 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data" (OuterVolumeSpecName: "config-data") pod "228bcf93-2f3d-461a-a006-b5429740bf6d" (UID: "228bcf93-2f3d-461a-a006-b5429740bf6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.005353 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "228bcf93-2f3d-461a-a006-b5429740bf6d" (UID: "228bcf93-2f3d-461a-a006-b5429740bf6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.044388 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.044433 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.044446 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.044456 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.044466 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wc4l\" (UniqueName: \"kubernetes.io/projected/228bcf93-2f3d-461a-a006-b5429740bf6d-kube-api-access-9wc4l\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.044479 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/228bcf93-2f3d-461a-a006-b5429740bf6d-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.285183 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.405045 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"228bcf93-2f3d-461a-a006-b5429740bf6d","Type":"ContainerDied","Data":"cdecde943d49e772ab686d40503745d965153b26ea15d65028de147043455bb7"} Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.405274 4894 scope.go:117] "RemoveContainer" containerID="c2c50ad9b4115480c242cbd5f35d9418373a3cb54dbd904a4bf68dc7a14accad" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.405402 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.411569 4894 generic.go:334] "Generic (PLEG): container finished" podID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerID="fcd772eee6a7ebb5c88934d022ebd48ecddd199bbee9c9360156d698935d92e1" exitCode=0 Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.411627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"38368ffb-25fe-478c-a1ab-5ca2b4a0c376","Type":"ContainerDied","Data":"fcd772eee6a7ebb5c88934d022ebd48ecddd199bbee9c9360156d698935d92e1"} Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.411651 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"38368ffb-25fe-478c-a1ab-5ca2b4a0c376","Type":"ContainerDied","Data":"40da522b4d68dc127a9e14be52191ea63f96a20ddb4b1e6d196cd8a1c5ff5d29"} Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.411662 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40da522b4d68dc127a9e14be52191ea63f96a20ddb4b1e6d196cd8a1c5ff5d29" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.413513 4894 generic.go:334] "Generic (PLEG): container finished" podID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerID="7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c" exitCode=0 Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.413757 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.415716 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.416282 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8f81cda0-aeed-41c1-9941-5f1ba92a88c5","Type":"ContainerDied","Data":"7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c"} Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.416329 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8f81cda0-aeed-41c1-9941-5f1ba92a88c5","Type":"ContainerDied","Data":"bb7c046121b938465f7c400dcff686e5d628a64787e9a627274c2bf17d2e9ea3"} Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.452786 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-plugins-conf\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.452832 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-erlang-cookie\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.452921 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-pod-info\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.452947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2zbf\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-kube-api-access-d2zbf\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.452979 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-config-data\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453452 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-tls\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453578 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-confd\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453630 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-server-conf\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453649 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-plugins\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453707 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-erlang-cookie-secret\") pod \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\" (UID: \"8f81cda0-aeed-41c1-9941-5f1ba92a88c5\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.453636 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.454202 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.454545 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.465468 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-kube-api-access-d2zbf" (OuterVolumeSpecName: "kube-api-access-d2zbf") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "kube-api-access-d2zbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.469435 4894 scope.go:117] "RemoveContainer" containerID="7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.470095 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-pod-info" (OuterVolumeSpecName: "pod-info") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.470385 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.474422 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.493257 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856" (OuterVolumeSpecName: "persistence") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "pvc-a02fe98a-a03c-467e-9da3-9212710ec856". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.501202 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.522574 4894 scope.go:117] "RemoveContainer" containerID="f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.535978 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-config-data" (OuterVolumeSpecName: "config-data") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.538006 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559353 4894 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559387 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2zbf\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-kube-api-access-d2zbf\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559398 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559425 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") on node \"crc\" " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559434 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559444 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559453 4894 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559462 4894 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.559471 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.589409 4894 scope.go:117] "RemoveContainer" containerID="7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.589861 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c\": container with ID starting with 7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c not found: ID does not exist" containerID="7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.590009 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c"} err="failed to get container status \"7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c\": rpc error: code = NotFound desc = could not find container \"7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c\": container with ID starting with 7a2156ac37ac04cf0090f6432ecaa94288392c1139807c208bf09c997b1e327c not found: ID does not exist" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.590081 4894 scope.go:117] "RemoveContainer" containerID="f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.590548 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e\": container with ID starting with f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e not found: ID does not exist" containerID="f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.590615 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e"} err="failed to get container status \"f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e\": rpc error: code = NotFound desc = could not find container \"f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e\": container with ID starting with f96ff8dbfe0ac635bef1ee6ea7e98c6460c6193da58e090aa1b2ac1298957b1e not found: ID does not exist" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.599971 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-server-conf" (OuterVolumeSpecName: "server-conf") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.622529 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.622572 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.648700 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.660794 4894 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.661118 4894 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a02fe98a-a03c-467e-9da3-9212710ec856" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856") on node "crc" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.662475 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-logs\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.662580 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-scripts\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.662816 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-internal-tls-certs\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.663014 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.663057 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data-custom\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.663093 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdftv\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-kube-api-access-cdftv\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.663141 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.663312 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-combined-ca-bundle\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.663412 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-certs\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.664142 4894 reconciler_common.go:293] "Volume detached for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.664181 4894 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.665779 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-logs" (OuterVolumeSpecName: "logs") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.668824 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-scripts" (OuterVolumeSpecName: "scripts") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.672524 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-kube-api-access-cdftv" (OuterVolumeSpecName: "kube-api-access-cdftv") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "kube-api-access-cdftv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.673305 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-certs" (OuterVolumeSpecName: "certs") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675351 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675805 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="setup-container" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675819 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="setup-container" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675867 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="rabbitmq" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675875 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="rabbitmq" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675888 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81df0908-03a5-439b-858d-e6e78ca7cce1" containerName="cloudkitty-storageinit" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675893 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="81df0908-03a5-439b-858d-e6e78ca7cce1" containerName="cloudkitty-storageinit" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675904 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="rabbitmq" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675909 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="rabbitmq" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675931 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api-log" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675937 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api-log" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675948 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675955 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.675965 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="setup-container" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.675970 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="setup-container" Dec 10 19:20:37 crc kubenswrapper[4894]: E1210 19:20:37.676006 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228bcf93-2f3d-461a-a006-b5429740bf6d" containerName="cloudkitty-proc" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676014 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="228bcf93-2f3d-461a-a006-b5429740bf6d" containerName="cloudkitty-proc" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676453 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" containerName="rabbitmq" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676478 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="228bcf93-2f3d-461a-a006-b5429740bf6d" containerName="cloudkitty-proc" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676489 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676498 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" containerName="cloudkitty-api-log" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676511 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" containerName="rabbitmq" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.676545 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="81df0908-03a5-439b-858d-e6e78ca7cce1" containerName="cloudkitty-storageinit" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.677304 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.678386 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.680060 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.703056 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data" (OuterVolumeSpecName: "config-data") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.703773 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.705805 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.707366 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.707655 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.708236 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.708418 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.709084 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.709214 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7xk4b" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.715723 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.721416 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.726906 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.733780 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.736349 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8f81cda0-aeed-41c1-9941-5f1ba92a88c5" (UID: "8f81cda0-aeed-41c1-9941-5f1ba92a88c5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.751157 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.766190 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.766268 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs\") pod \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\" (UID: \"38368ffb-25fe-478c-a1ab-5ca2b4a0c376\") " Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767126 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767177 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767187 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767195 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767204 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8f81cda0-aeed-41c1-9941-5f1ba92a88c5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767211 4894 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767219 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767229 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdftv\" (UniqueName: \"kubernetes.io/projected/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-kube-api-access-cdftv\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767246 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: W1210 19:20:37.767409 4894 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/38368ffb-25fe-478c-a1ab-5ca2b4a0c376/volumes/kubernetes.io~secret/public-tls-certs Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.767422 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "38368ffb-25fe-478c-a1ab-5ca2b4a0c376" (UID: "38368ffb-25fe-478c-a1ab-5ca2b4a0c376"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.868659 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-config-data\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869073 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869096 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-scripts\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869111 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869126 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869230 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869255 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869350 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869409 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869442 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv7x7\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-kube-api-access-wv7x7\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd2wv\" (UniqueName: \"kubernetes.io/projected/26836cdd-cd71-4430-9a64-bfd4cfa982c6-kube-api-access-bd2wv\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/26836cdd-cd71-4430-9a64-bfd4cfa982c6-certs\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869537 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869574 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.869636 4894 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38368ffb-25fe-478c-a1ab-5ca2b4a0c376-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971264 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971327 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971390 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971408 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv7x7\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-kube-api-access-wv7x7\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd2wv\" (UniqueName: \"kubernetes.io/projected/26836cdd-cd71-4430-9a64-bfd4cfa982c6-kube-api-access-bd2wv\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/26836cdd-cd71-4430-9a64-bfd4cfa982c6-certs\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971469 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971497 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971528 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-config-data\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971577 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971594 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-scripts\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971610 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971626 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971650 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971671 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.971685 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.972468 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.974443 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.974816 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.977630 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.978531 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.980264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-config-data\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.981146 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.981301 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.983100 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.983131 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c975b249a0e4a5d819d74290b52a518dd30d8c0b64914f36ad533f66100f6cb7/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 10 19:20:37 crc kubenswrapper[4894]: I1210 19:20:37.989282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-scripts\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.008759 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/26836cdd-cd71-4430-9a64-bfd4cfa982c6-certs\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.009382 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.009540 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/26836cdd-cd71-4430-9a64-bfd4cfa982c6-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.009939 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.020712 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv7x7\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-kube-api-access-wv7x7\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.037344 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d27beb4-0b6e-4543-b3a7-153fc30e2fe4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.037962 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd2wv\" (UniqueName: \"kubernetes.io/projected/26836cdd-cd71-4430-9a64-bfd4cfa982c6-kube-api-access-bd2wv\") pod \"cloudkitty-proc-0\" (UID: \"26836cdd-cd71-4430-9a64-bfd4cfa982c6\") " pod="openstack/cloudkitty-proc-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.127654 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.158315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d24fd592-c53c-42b1-8849-8a514156219b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d24fd592-c53c-42b1-8849-8a514156219b\") pod \"rabbitmq-server-0\" (UID: \"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4\") " pod="openstack/rabbitmq-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.179125 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.192692 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.196015 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.205801 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.208793 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.208999 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.209091 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.209099 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.209135 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.209295 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wnk5r" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.209424 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282471 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282519 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282537 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282561 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/800b4501-b71a-4854-88f9-4651f92b375e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282591 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282609 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282671 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2d4q\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-kube-api-access-q2d4q\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282712 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.282778 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/800b4501-b71a-4854-88f9-4651f92b375e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.295342 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.349352 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384701 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/800b4501-b71a-4854-88f9-4651f92b375e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384837 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384879 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384895 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/800b4501-b71a-4854-88f9-4651f92b375e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.384962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.385030 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2d4q\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-kube-api-access-q2d4q\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.385068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.385493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.385938 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.389048 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.393642 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.393673 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.398015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/800b4501-b71a-4854-88f9-4651f92b375e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.406391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/800b4501-b71a-4854-88f9-4651f92b375e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.406542 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.407412 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/800b4501-b71a-4854-88f9-4651f92b375e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.434717 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.434760 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8fde555e75f624d7793c9908446e92e60787eecb4e949c1f5f68bb115b06429c/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.435030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2d4q\" (UniqueName: \"kubernetes.io/projected/800b4501-b71a-4854-88f9-4651f92b375e-kube-api-access-q2d4q\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.604644 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10e3654c-3444-4fd9-9163-0b93a2089f15","Type":"ContainerStarted","Data":"8fc8ab873890ce628caf6d7d5141850943003e76475fb1212c6f80514db6e21b"} Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.605271 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.617957 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.759690 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a02fe98a-a03c-467e-9da3-9212710ec856\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a02fe98a-a03c-467e-9da3-9212710ec856\") pod \"rabbitmq-cell1-server-0\" (UID: \"800b4501-b71a-4854-88f9-4651f92b375e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.794911 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.820413953 podStartE2EDuration="9.79488624s" podCreationTimestamp="2025-12-10 19:20:29 +0000 UTC" firstStartedPulling="2025-12-10 19:20:30.105286118 +0000 UTC m=+1512.900133910" lastFinishedPulling="2025-12-10 19:20:37.079758405 +0000 UTC m=+1519.874606197" observedRunningTime="2025-12-10 19:20:38.65552896 +0000 UTC m=+1521.450376752" watchObservedRunningTime="2025-12-10 19:20:38.79488624 +0000 UTC m=+1521.589734042" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.796921 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.818218 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.832137 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.834779 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.839177 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.839341 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.839381 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.844367 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.854141 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896247 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896321 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-scripts\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896350 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e167ef7-40fe-48c3-92b8-670eff58b203-logs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896423 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896445 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e167ef7-40fe-48c3-92b8-670eff58b203-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896473 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-config-data\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:38 crc kubenswrapper[4894]: I1210 19:20:38.896486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc7hl\" (UniqueName: \"kubernetes.io/projected/5e167ef7-40fe-48c3-92b8-670eff58b203-kube-api-access-vc7hl\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.000525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.000820 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e167ef7-40fe-48c3-92b8-670eff58b203-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.000907 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-config-data\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.000925 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc7hl\" (UniqueName: \"kubernetes.io/projected/5e167ef7-40fe-48c3-92b8-670eff58b203-kube-api-access-vc7hl\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.001021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.001045 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.001070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-scripts\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.001102 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.001152 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e167ef7-40fe-48c3-92b8-670eff58b203-logs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.001545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e167ef7-40fe-48c3-92b8-670eff58b203-logs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.009441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/5e167ef7-40fe-48c3-92b8-670eff58b203-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.009628 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.010043 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.012445 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.013087 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.015191 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-scripts\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.017684 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e167ef7-40fe-48c3-92b8-670eff58b203-config-data\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.030540 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc7hl\" (UniqueName: \"kubernetes.io/projected/5e167ef7-40fe-48c3-92b8-670eff58b203-kube-api-access-vc7hl\") pod \"cloudkitty-api-0\" (UID: \"5e167ef7-40fe-48c3-92b8-670eff58b203\") " pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.158157 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-595979776c-g9qpm"] Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.162945 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.167240 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.187099 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-595979776c-g9qpm"] Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.190586 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.226447 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-config\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310529 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-nb\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310596 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-sb\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310717 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-svc\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310739 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvpkj\" (UniqueName: \"kubernetes.io/projected/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-kube-api-access-kvpkj\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-openstack-edpm-ipam\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.310826 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-swift-storage-0\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.379865 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412550 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-svc\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvpkj\" (UniqueName: \"kubernetes.io/projected/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-kube-api-access-kvpkj\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412646 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-openstack-edpm-ipam\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412686 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-swift-storage-0\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412757 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-config\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-nb\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.412863 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-sb\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.413542 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-svc\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.413707 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-sb\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.414493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-swift-storage-0\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.414592 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-config\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.414624 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-nb\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.421390 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-openstack-edpm-ipam\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.434545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvpkj\" (UniqueName: \"kubernetes.io/projected/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-kube-api-access-kvpkj\") pod \"dnsmasq-dns-595979776c-g9qpm\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.499077 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.564385 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="228bcf93-2f3d-461a-a006-b5429740bf6d" path="/var/lib/kubelet/pods/228bcf93-2f3d-461a-a006-b5429740bf6d/volumes" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.565268 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38368ffb-25fe-478c-a1ab-5ca2b4a0c376" path="/var/lib/kubelet/pods/38368ffb-25fe-478c-a1ab-5ca2b4a0c376/volumes" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.566023 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f81cda0-aeed-41c1-9941-5f1ba92a88c5" path="/var/lib/kubelet/pods/8f81cda0-aeed-41c1-9941-5f1ba92a88c5/volumes" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.567474 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97355ddc-532e-46a2-a568-31f4a5e07030" path="/var/lib/kubelet/pods/97355ddc-532e-46a2-a568-31f4a5e07030/volumes" Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.612312 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.645186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4","Type":"ContainerStarted","Data":"94824fdf1fd05785502921b33ca3f21433edeb9d0df14bb0bc67e3e8abecfb55"} Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.668874 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"26836cdd-cd71-4430-9a64-bfd4cfa982c6","Type":"ContainerStarted","Data":"653d44ba8eac93ec73596aac87e71ea5a6b2b048313244f3f86ef8389d5f9575"} Dec 10 19:20:39 crc kubenswrapper[4894]: I1210 19:20:39.784044 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 10 19:20:39 crc kubenswrapper[4894]: W1210 19:20:39.787161 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e167ef7_40fe_48c3_92b8_670eff58b203.slice/crio-8e781257ca1dc7904e48f4bbc4b7c99020e58fcf80c855e3bbc65749c104152e WatchSource:0}: Error finding container 8e781257ca1dc7904e48f4bbc4b7c99020e58fcf80c855e3bbc65749c104152e: Status 404 returned error can't find the container with id 8e781257ca1dc7904e48f4bbc4b7c99020e58fcf80c855e3bbc65749c104152e Dec 10 19:20:40 crc kubenswrapper[4894]: W1210 19:20:40.009963 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90bd2edb_9b2e_40fc_8d45_7167fcf3cc4c.slice/crio-aa83d66a0c5f674d9cb10fa9deee385f6e2f49b7b8e1f24972ce47d0d6d067f1 WatchSource:0}: Error finding container aa83d66a0c5f674d9cb10fa9deee385f6e2f49b7b8e1f24972ce47d0d6d067f1: Status 404 returned error can't find the container with id aa83d66a0c5f674d9cb10fa9deee385f6e2f49b7b8e1f24972ce47d0d6d067f1 Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.010060 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-595979776c-g9qpm"] Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.683069 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"26836cdd-cd71-4430-9a64-bfd4cfa982c6","Type":"ContainerStarted","Data":"8706819b417bfb3281ea760b22ebfe196eb73be71b380d4a7ad62a13e4944c63"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.684050 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"800b4501-b71a-4854-88f9-4651f92b375e","Type":"ContainerStarted","Data":"a1cb52f785aa40f8d1fff0f17f5816a68f30c5cd8ffefe755164f95ffa4792c3"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.685759 4894 generic.go:334] "Generic (PLEG): container finished" podID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerID="8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5" exitCode=0 Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.685798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595979776c-g9qpm" event={"ID":"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c","Type":"ContainerDied","Data":"8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.685813 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595979776c-g9qpm" event={"ID":"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c","Type":"ContainerStarted","Data":"aa83d66a0c5f674d9cb10fa9deee385f6e2f49b7b8e1f24972ce47d0d6d067f1"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.711014 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"5e167ef7-40fe-48c3-92b8-670eff58b203","Type":"ContainerStarted","Data":"9ee85c8503ae6dd999da047922a4ed5456df4beb8d9726d70656af1a38a1dad9"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.711265 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"5e167ef7-40fe-48c3-92b8-670eff58b203","Type":"ContainerStarted","Data":"41476c3fa2c5bc68a86c209c3d45f5e228017083b2411de39d990dcfda64142f"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.711339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"5e167ef7-40fe-48c3-92b8-670eff58b203","Type":"ContainerStarted","Data":"8e781257ca1dc7904e48f4bbc4b7c99020e58fcf80c855e3bbc65749c104152e"} Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.712216 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.723396 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.487596125 podStartE2EDuration="3.723376088s" podCreationTimestamp="2025-12-10 19:20:37 +0000 UTC" firstStartedPulling="2025-12-10 19:20:39.239018537 +0000 UTC m=+1522.033866329" lastFinishedPulling="2025-12-10 19:20:39.4747985 +0000 UTC m=+1522.269646292" observedRunningTime="2025-12-10 19:20:40.710275576 +0000 UTC m=+1523.505123378" watchObservedRunningTime="2025-12-10 19:20:40.723376088 +0000 UTC m=+1523.518223880" Dec 10 19:20:40 crc kubenswrapper[4894]: I1210 19:20:40.816123 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.816102869 podStartE2EDuration="2.816102869s" podCreationTimestamp="2025-12-10 19:20:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:40.8056181 +0000 UTC m=+1523.600465892" watchObservedRunningTime="2025-12-10 19:20:40.816102869 +0000 UTC m=+1523.610950661" Dec 10 19:20:41 crc kubenswrapper[4894]: I1210 19:20:41.722303 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"800b4501-b71a-4854-88f9-4651f92b375e","Type":"ContainerStarted","Data":"11a0b029964647fa3af5eb265a116b77d3ec15feeaf392c9f643080b302b1f50"} Dec 10 19:20:41 crc kubenswrapper[4894]: I1210 19:20:41.724646 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595979776c-g9qpm" event={"ID":"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c","Type":"ContainerStarted","Data":"a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd"} Dec 10 19:20:41 crc kubenswrapper[4894]: I1210 19:20:41.724778 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:41 crc kubenswrapper[4894]: I1210 19:20:41.727680 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4","Type":"ContainerStarted","Data":"436bab3926244f9ae4fc406d6efc0971c18201a2d206a53ce3ed156c04ba432c"} Dec 10 19:20:41 crc kubenswrapper[4894]: I1210 19:20:41.850770 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-595979776c-g9qpm" podStartSLOduration=2.850755618 podStartE2EDuration="2.850755618s" podCreationTimestamp="2025-12-10 19:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:41.850220273 +0000 UTC m=+1524.645068075" watchObservedRunningTime="2025-12-10 19:20:41.850755618 +0000 UTC m=+1524.645603400" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.501733 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.600256 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78468d7767-8ppb5"] Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.600496 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerName="dnsmasq-dns" containerID="cri-o://3742f7be0244d31b9994feb14d1b92fc6e804b562b8507d90ebbb61a2a977c8e" gracePeriod=10 Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.764127 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5475ccd585-g6mjv"] Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.767114 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.785874 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5475ccd585-g6mjv"] Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.834782 4894 generic.go:334] "Generic (PLEG): container finished" podID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerID="3742f7be0244d31b9994feb14d1b92fc6e804b562b8507d90ebbb61a2a977c8e" exitCode=0 Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.834874 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" event={"ID":"f46b5f26-4bcb-4796-9a8b-706735947bc9","Type":"ContainerDied","Data":"3742f7be0244d31b9994feb14d1b92fc6e804b562b8507d90ebbb61a2a977c8e"} Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.869731 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-dns-swift-storage-0\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.869818 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-ovsdbserver-sb\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.869919 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-openstack-edpm-ipam\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.869969 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-ovsdbserver-nb\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.870038 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-dns-svc\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.870098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pp9d\" (UniqueName: \"kubernetes.io/projected/505346a3-228e-4aa5-8380-b057187ca6d0-kube-api-access-9pp9d\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.870148 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-config\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.973517 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-dns-svc\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.973855 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pp9d\" (UniqueName: \"kubernetes.io/projected/505346a3-228e-4aa5-8380-b057187ca6d0-kube-api-access-9pp9d\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.973984 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-config\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.974016 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-dns-swift-storage-0\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.974059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-ovsdbserver-sb\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.974118 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-openstack-edpm-ipam\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.974151 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-ovsdbserver-nb\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.974579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-dns-svc\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.974989 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-ovsdbserver-nb\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.975636 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-dns-swift-storage-0\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.975766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-openstack-edpm-ipam\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.975882 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-config\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:49 crc kubenswrapper[4894]: I1210 19:20:49.976013 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/505346a3-228e-4aa5-8380-b057187ca6d0-ovsdbserver-sb\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.007898 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pp9d\" (UniqueName: \"kubernetes.io/projected/505346a3-228e-4aa5-8380-b057187ca6d0-kube-api-access-9pp9d\") pod \"dnsmasq-dns-5475ccd585-g6mjv\" (UID: \"505346a3-228e-4aa5-8380-b057187ca6d0\") " pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.088741 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.251273 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.282625 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-config\") pod \"f46b5f26-4bcb-4796-9a8b-706735947bc9\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.282784 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6r9z\" (UniqueName: \"kubernetes.io/projected/f46b5f26-4bcb-4796-9a8b-706735947bc9-kube-api-access-f6r9z\") pod \"f46b5f26-4bcb-4796-9a8b-706735947bc9\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.282820 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-sb\") pod \"f46b5f26-4bcb-4796-9a8b-706735947bc9\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.282920 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-nb\") pod \"f46b5f26-4bcb-4796-9a8b-706735947bc9\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.282962 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-swift-storage-0\") pod \"f46b5f26-4bcb-4796-9a8b-706735947bc9\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.283026 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-svc\") pod \"f46b5f26-4bcb-4796-9a8b-706735947bc9\" (UID: \"f46b5f26-4bcb-4796-9a8b-706735947bc9\") " Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.308513 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f46b5f26-4bcb-4796-9a8b-706735947bc9-kube-api-access-f6r9z" (OuterVolumeSpecName: "kube-api-access-f6r9z") pod "f46b5f26-4bcb-4796-9a8b-706735947bc9" (UID: "f46b5f26-4bcb-4796-9a8b-706735947bc9"). InnerVolumeSpecName "kube-api-access-f6r9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.347461 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f46b5f26-4bcb-4796-9a8b-706735947bc9" (UID: "f46b5f26-4bcb-4796-9a8b-706735947bc9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.369869 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-config" (OuterVolumeSpecName: "config") pod "f46b5f26-4bcb-4796-9a8b-706735947bc9" (UID: "f46b5f26-4bcb-4796-9a8b-706735947bc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.382559 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f46b5f26-4bcb-4796-9a8b-706735947bc9" (UID: "f46b5f26-4bcb-4796-9a8b-706735947bc9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.387434 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.387469 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6r9z\" (UniqueName: \"kubernetes.io/projected/f46b5f26-4bcb-4796-9a8b-706735947bc9-kube-api-access-f6r9z\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.387482 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.387492 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.395602 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f46b5f26-4bcb-4796-9a8b-706735947bc9" (UID: "f46b5f26-4bcb-4796-9a8b-706735947bc9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.398476 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f46b5f26-4bcb-4796-9a8b-706735947bc9" (UID: "f46b5f26-4bcb-4796-9a8b-706735947bc9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.489136 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.489185 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f46b5f26-4bcb-4796-9a8b-706735947bc9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.630756 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5475ccd585-g6mjv"] Dec 10 19:20:50 crc kubenswrapper[4894]: W1210 19:20:50.641884 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod505346a3_228e_4aa5_8380_b057187ca6d0.slice/crio-ee58b283ac77ebb1fa4c365c8f546074af15f8cb083ea23cf9ad433c255e8729 WatchSource:0}: Error finding container ee58b283ac77ebb1fa4c365c8f546074af15f8cb083ea23cf9ad433c255e8729: Status 404 returned error can't find the container with id ee58b283ac77ebb1fa4c365c8f546074af15f8cb083ea23cf9ad433c255e8729 Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.853361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" event={"ID":"505346a3-228e-4aa5-8380-b057187ca6d0","Type":"ContainerStarted","Data":"ee58b283ac77ebb1fa4c365c8f546074af15f8cb083ea23cf9ad433c255e8729"} Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.875114 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" event={"ID":"f46b5f26-4bcb-4796-9a8b-706735947bc9","Type":"ContainerDied","Data":"a651524e520c39506bce38ecd09a1c0f7119bc9c7b3c4e93dd6425d0e64e80a5"} Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.875187 4894 scope.go:117] "RemoveContainer" containerID="3742f7be0244d31b9994feb14d1b92fc6e804b562b8507d90ebbb61a2a977c8e" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.875199 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78468d7767-8ppb5" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.932181 4894 scope.go:117] "RemoveContainer" containerID="d194c2132d2625084689fc37912f640205a1e5d3f84e476b25ba84dcfd211378" Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.960869 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78468d7767-8ppb5"] Dec 10 19:20:50 crc kubenswrapper[4894]: I1210 19:20:50.972595 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78468d7767-8ppb5"] Dec 10 19:20:51 crc kubenswrapper[4894]: I1210 19:20:51.552991 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" path="/var/lib/kubelet/pods/f46b5f26-4bcb-4796-9a8b-706735947bc9/volumes" Dec 10 19:20:51 crc kubenswrapper[4894]: I1210 19:20:51.888933 4894 generic.go:334] "Generic (PLEG): container finished" podID="505346a3-228e-4aa5-8380-b057187ca6d0" containerID="b68f0a278b41fb262275a655348979c5863818458039a46fdb5137ddd77795bf" exitCode=0 Dec 10 19:20:51 crc kubenswrapper[4894]: I1210 19:20:51.888970 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" event={"ID":"505346a3-228e-4aa5-8380-b057187ca6d0","Type":"ContainerDied","Data":"b68f0a278b41fb262275a655348979c5863818458039a46fdb5137ddd77795bf"} Dec 10 19:20:52 crc kubenswrapper[4894]: I1210 19:20:52.902118 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" event={"ID":"505346a3-228e-4aa5-8380-b057187ca6d0","Type":"ContainerStarted","Data":"f733671871d74b4de6bbb31f3d08fd2344f733779c4ac13152ae1901e4e9d0ee"} Dec 10 19:20:52 crc kubenswrapper[4894]: I1210 19:20:52.903409 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:20:52 crc kubenswrapper[4894]: I1210 19:20:52.938286 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" podStartSLOduration=3.938256521 podStartE2EDuration="3.938256521s" podCreationTimestamp="2025-12-10 19:20:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:52.929690074 +0000 UTC m=+1535.724537876" watchObservedRunningTime="2025-12-10 19:20:52.938256521 +0000 UTC m=+1535.733104323" Dec 10 19:20:59 crc kubenswrapper[4894]: I1210 19:20:59.618937 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.091132 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5475ccd585-g6mjv" Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.161297 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-595979776c-g9qpm"] Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.161530 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-595979776c-g9qpm" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerName="dnsmasq-dns" containerID="cri-o://a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd" gracePeriod=10 Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.833287 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.995526 4894 generic.go:334] "Generic (PLEG): container finished" podID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerID="a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd" exitCode=0 Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.995594 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-595979776c-g9qpm" Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.995605 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595979776c-g9qpm" event={"ID":"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c","Type":"ContainerDied","Data":"a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd"} Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.995661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-595979776c-g9qpm" event={"ID":"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c","Type":"ContainerDied","Data":"aa83d66a0c5f674d9cb10fa9deee385f6e2f49b7b8e1f24972ce47d0d6d067f1"} Dec 10 19:21:00 crc kubenswrapper[4894]: I1210 19:21:00.995682 4894 scope.go:117] "RemoveContainer" containerID="a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.015616 4894 scope.go:117] "RemoveContainer" containerID="8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.032760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-svc\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.032875 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-config\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.032921 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-swift-storage-0\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.033061 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-openstack-edpm-ipam\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.033135 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvpkj\" (UniqueName: \"kubernetes.io/projected/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-kube-api-access-kvpkj\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.033175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-sb\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.033211 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-nb\") pod \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\" (UID: \"90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c\") " Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.038790 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-kube-api-access-kvpkj" (OuterVolumeSpecName: "kube-api-access-kvpkj") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "kube-api-access-kvpkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.039304 4894 scope.go:117] "RemoveContainer" containerID="a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd" Dec 10 19:21:01 crc kubenswrapper[4894]: E1210 19:21:01.039752 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd\": container with ID starting with a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd not found: ID does not exist" containerID="a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.039786 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd"} err="failed to get container status \"a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd\": rpc error: code = NotFound desc = could not find container \"a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd\": container with ID starting with a5fd1fe063bc55a01c0dfbc409d8983624e9678bce69f8676f4783bb61ac3dfd not found: ID does not exist" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.039811 4894 scope.go:117] "RemoveContainer" containerID="8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5" Dec 10 19:21:01 crc kubenswrapper[4894]: E1210 19:21:01.040049 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5\": container with ID starting with 8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5 not found: ID does not exist" containerID="8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.040090 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5"} err="failed to get container status \"8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5\": rpc error: code = NotFound desc = could not find container \"8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5\": container with ID starting with 8a32e146aba36ab5b2261ea7e1afb94e25c995a8fa5780258e8112cd57cc3dc5 not found: ID does not exist" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.095368 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.099122 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.099158 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.103408 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.104072 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-config" (OuterVolumeSpecName: "config") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.107286 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" (UID: "90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135477 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135508 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135520 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvpkj\" (UniqueName: \"kubernetes.io/projected/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-kube-api-access-kvpkj\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135532 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135540 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135550 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.135559 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.341346 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-595979776c-g9qpm"] Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.351721 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-595979776c-g9qpm"] Dec 10 19:21:01 crc kubenswrapper[4894]: I1210 19:21:01.553664 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" path="/var/lib/kubelet/pods/90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c/volumes" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.472665 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47"] Dec 10 19:21:12 crc kubenswrapper[4894]: E1210 19:21:12.473602 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerName="dnsmasq-dns" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.473614 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerName="dnsmasq-dns" Dec 10 19:21:12 crc kubenswrapper[4894]: E1210 19:21:12.473622 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerName="init" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.473629 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerName="init" Dec 10 19:21:12 crc kubenswrapper[4894]: E1210 19:21:12.473643 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerName="dnsmasq-dns" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.473649 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerName="dnsmasq-dns" Dec 10 19:21:12 crc kubenswrapper[4894]: E1210 19:21:12.473657 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerName="init" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.473663 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerName="init" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.473918 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f46b5f26-4bcb-4796-9a8b-706735947bc9" containerName="dnsmasq-dns" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.473955 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="90bd2edb-9b2e-40fc-8d45-7167fcf3cc4c" containerName="dnsmasq-dns" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.474608 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.478726 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.478986 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.481553 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.482607 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.508590 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47"] Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.580299 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkr7g\" (UniqueName: \"kubernetes.io/projected/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-kube-api-access-bkr7g\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.580371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.580424 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.580455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.681735 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkr7g\" (UniqueName: \"kubernetes.io/projected/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-kube-api-access-bkr7g\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.681811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.681853 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.681890 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.690680 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.691098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.695535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.710554 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkr7g\" (UniqueName: \"kubernetes.io/projected/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-kube-api-access-bkr7g\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lth47\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:12 crc kubenswrapper[4894]: I1210 19:21:12.796441 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.144805 4894 generic.go:334] "Generic (PLEG): container finished" podID="3d27beb4-0b6e-4543-b3a7-153fc30e2fe4" containerID="436bab3926244f9ae4fc406d6efc0971c18201a2d206a53ce3ed156c04ba432c" exitCode=0 Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.145068 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4","Type":"ContainerDied","Data":"436bab3926244f9ae4fc406d6efc0971c18201a2d206a53ce3ed156c04ba432c"} Dec 10 19:21:13 crc kubenswrapper[4894]: W1210 19:21:13.444534 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod919d6c2a_def8_4ef2_b1b0_9f65d61be6fe.slice/crio-e5dcb24e820754e1af506501afebe4bff7b0f6a5a1097ed818df672d66ee733b WatchSource:0}: Error finding container e5dcb24e820754e1af506501afebe4bff7b0f6a5a1097ed818df672d66ee733b: Status 404 returned error can't find the container with id e5dcb24e820754e1af506501afebe4bff7b0f6a5a1097ed818df672d66ee733b Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.446003 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47"] Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.768470 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-czlzc"] Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.770561 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.792038 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-czlzc"] Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.907715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-utilities\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.907988 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-catalog-content\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:13 crc kubenswrapper[4894]: I1210 19:21:13.908367 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjs4c\" (UniqueName: \"kubernetes.io/projected/1d18d063-dbd5-41bc-88a5-db817a1b75bb-kube-api-access-rjs4c\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.010128 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjs4c\" (UniqueName: \"kubernetes.io/projected/1d18d063-dbd5-41bc-88a5-db817a1b75bb-kube-api-access-rjs4c\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.010468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-utilities\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.010523 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-catalog-content\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.010989 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-catalog-content\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.011154 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-utilities\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.031541 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjs4c\" (UniqueName: \"kubernetes.io/projected/1d18d063-dbd5-41bc-88a5-db817a1b75bb-kube-api-access-rjs4c\") pod \"certified-operators-czlzc\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.093710 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.177699 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" event={"ID":"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe","Type":"ContainerStarted","Data":"e5dcb24e820754e1af506501afebe4bff7b0f6a5a1097ed818df672d66ee733b"} Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.180665 4894 generic.go:334] "Generic (PLEG): container finished" podID="800b4501-b71a-4854-88f9-4651f92b375e" containerID="11a0b029964647fa3af5eb265a116b77d3ec15feeaf392c9f643080b302b1f50" exitCode=0 Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.180728 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"800b4501-b71a-4854-88f9-4651f92b375e","Type":"ContainerDied","Data":"11a0b029964647fa3af5eb265a116b77d3ec15feeaf392c9f643080b302b1f50"} Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.208544 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d27beb4-0b6e-4543-b3a7-153fc30e2fe4","Type":"ContainerStarted","Data":"f1e982e72d7cd38f798d5dab8c2c8140dd85f3fa7d9572068d3816ed7b0f2714"} Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.209436 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.303839 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.303811845 podStartE2EDuration="37.303811845s" podCreationTimestamp="2025-12-10 19:20:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:14.273174018 +0000 UTC m=+1557.068021830" watchObservedRunningTime="2025-12-10 19:21:14.303811845 +0000 UTC m=+1557.098659637" Dec 10 19:21:14 crc kubenswrapper[4894]: W1210 19:21:14.695936 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d18d063_dbd5_41bc_88a5_db817a1b75bb.slice/crio-205ea045d87212c96217901987a693a578c129ed38ad4735852e7c8e77304370 WatchSource:0}: Error finding container 205ea045d87212c96217901987a693a578c129ed38ad4735852e7c8e77304370: Status 404 returned error can't find the container with id 205ea045d87212c96217901987a693a578c129ed38ad4735852e7c8e77304370 Dec 10 19:21:14 crc kubenswrapper[4894]: I1210 19:21:14.727188 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-czlzc"] Dec 10 19:21:15 crc kubenswrapper[4894]: I1210 19:21:15.221335 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"800b4501-b71a-4854-88f9-4651f92b375e","Type":"ContainerStarted","Data":"f563e4078bc39af1874517b34bf40bb79842a9f370f919e8dec23af00d2275cf"} Dec 10 19:21:15 crc kubenswrapper[4894]: I1210 19:21:15.223432 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:21:15 crc kubenswrapper[4894]: I1210 19:21:15.226073 4894 generic.go:334] "Generic (PLEG): container finished" podID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerID="bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f" exitCode=0 Dec 10 19:21:15 crc kubenswrapper[4894]: I1210 19:21:15.227038 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerDied","Data":"bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f"} Dec 10 19:21:15 crc kubenswrapper[4894]: I1210 19:21:15.227117 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerStarted","Data":"205ea045d87212c96217901987a693a578c129ed38ad4735852e7c8e77304370"} Dec 10 19:21:15 crc kubenswrapper[4894]: I1210 19:21:15.252028 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.252010435 podStartE2EDuration="37.252010435s" podCreationTimestamp="2025-12-10 19:20:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:15.247373078 +0000 UTC m=+1558.042220870" watchObservedRunningTime="2025-12-10 19:21:15.252010435 +0000 UTC m=+1558.046858227" Dec 10 19:21:17 crc kubenswrapper[4894]: I1210 19:21:17.081461 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.319029 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b8rsv"] Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.322382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.330436 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8rsv"] Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.356098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-utilities\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.356396 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-catalog-content\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.356566 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvjzp\" (UniqueName: \"kubernetes.io/projected/3d8ff699-e886-4443-b13d-257df5ff62f0-kube-api-access-wvjzp\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.458234 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-catalog-content\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.458706 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-catalog-content\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.458776 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvjzp\" (UniqueName: \"kubernetes.io/projected/3d8ff699-e886-4443-b13d-257df5ff62f0-kube-api-access-wvjzp\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.459001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-utilities\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.459436 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-utilities\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.477509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvjzp\" (UniqueName: \"kubernetes.io/projected/3d8ff699-e886-4443-b13d-257df5ff62f0-kube-api-access-wvjzp\") pod \"community-operators-b8rsv\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:18 crc kubenswrapper[4894]: I1210 19:21:18.650558 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:19 crc kubenswrapper[4894]: I1210 19:21:19.256506 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8rsv"] Dec 10 19:21:19 crc kubenswrapper[4894]: W1210 19:21:19.272013 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d8ff699_e886_4443_b13d_257df5ff62f0.slice/crio-bc77e3a1d3e6a07db24cc29e3afd5d2c297e85a248454937d44e367627033741 WatchSource:0}: Error finding container bc77e3a1d3e6a07db24cc29e3afd5d2c297e85a248454937d44e367627033741: Status 404 returned error can't find the container with id bc77e3a1d3e6a07db24cc29e3afd5d2c297e85a248454937d44e367627033741 Dec 10 19:21:19 crc kubenswrapper[4894]: I1210 19:21:19.299567 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerStarted","Data":"bc77e3a1d3e6a07db24cc29e3afd5d2c297e85a248454937d44e367627033741"} Dec 10 19:21:19 crc kubenswrapper[4894]: I1210 19:21:19.301587 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerStarted","Data":"33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2"} Dec 10 19:21:19 crc kubenswrapper[4894]: I1210 19:21:19.537150 4894 scope.go:117] "RemoveContainer" containerID="cf236dedc39a8f67620519bcb9ba4e9cef45cc34cf202fb40082f8055745524a" Dec 10 19:21:19 crc kubenswrapper[4894]: I1210 19:21:19.577603 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:21:19 crc kubenswrapper[4894]: I1210 19:21:19.577664 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:21:20 crc kubenswrapper[4894]: I1210 19:21:20.337237 4894 generic.go:334] "Generic (PLEG): container finished" podID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerID="33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2" exitCode=0 Dec 10 19:21:20 crc kubenswrapper[4894]: I1210 19:21:20.337556 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerDied","Data":"33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2"} Dec 10 19:21:21 crc kubenswrapper[4894]: I1210 19:21:21.355577 4894 generic.go:334] "Generic (PLEG): container finished" podID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerID="9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f" exitCode=0 Dec 10 19:21:21 crc kubenswrapper[4894]: I1210 19:21:21.355624 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerDied","Data":"9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f"} Dec 10 19:21:27 crc kubenswrapper[4894]: I1210 19:21:27.739413 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:21:28 crc kubenswrapper[4894]: I1210 19:21:28.352300 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 19:21:28 crc kubenswrapper[4894]: I1210 19:21:28.470618 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerStarted","Data":"13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b"} Dec 10 19:21:28 crc kubenswrapper[4894]: I1210 19:21:28.486589 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" event={"ID":"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe","Type":"ContainerStarted","Data":"0ce2413aca8bf7581de470b73ffaf44bd9b6d1151c0665d86a225ee43d0b87c1"} Dec 10 19:21:28 crc kubenswrapper[4894]: I1210 19:21:28.520650 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-czlzc" podStartSLOduration=3.036943256 podStartE2EDuration="15.520631307s" podCreationTimestamp="2025-12-10 19:21:13 +0000 UTC" firstStartedPulling="2025-12-10 19:21:15.22793341 +0000 UTC m=+1558.022781202" lastFinishedPulling="2025-12-10 19:21:27.711621451 +0000 UTC m=+1570.506469253" observedRunningTime="2025-12-10 19:21:28.499963787 +0000 UTC m=+1571.294811589" watchObservedRunningTime="2025-12-10 19:21:28.520631307 +0000 UTC m=+1571.315479099" Dec 10 19:21:28 crc kubenswrapper[4894]: I1210 19:21:28.532320 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" podStartSLOduration=2.244156128 podStartE2EDuration="16.53230428s" podCreationTimestamp="2025-12-10 19:21:12 +0000 UTC" firstStartedPulling="2025-12-10 19:21:13.447024439 +0000 UTC m=+1556.241872231" lastFinishedPulling="2025-12-10 19:21:27.735172581 +0000 UTC m=+1570.530020383" observedRunningTime="2025-12-10 19:21:28.52506314 +0000 UTC m=+1571.319910942" watchObservedRunningTime="2025-12-10 19:21:28.53230428 +0000 UTC m=+1571.327152082" Dec 10 19:21:28 crc kubenswrapper[4894]: I1210 19:21:28.849009 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:21:29 crc kubenswrapper[4894]: I1210 19:21:29.497826 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerStarted","Data":"5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e"} Dec 10 19:21:30 crc kubenswrapper[4894]: I1210 19:21:30.509395 4894 generic.go:334] "Generic (PLEG): container finished" podID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerID="5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e" exitCode=0 Dec 10 19:21:30 crc kubenswrapper[4894]: I1210 19:21:30.509447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerDied","Data":"5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e"} Dec 10 19:21:31 crc kubenswrapper[4894]: I1210 19:21:31.522619 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerStarted","Data":"c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435"} Dec 10 19:21:31 crc kubenswrapper[4894]: I1210 19:21:31.547886 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b8rsv" podStartSLOduration=10.229521837 podStartE2EDuration="13.547822624s" podCreationTimestamp="2025-12-10 19:21:18 +0000 UTC" firstStartedPulling="2025-12-10 19:21:27.647348566 +0000 UTC m=+1570.442196358" lastFinishedPulling="2025-12-10 19:21:30.965649353 +0000 UTC m=+1573.760497145" observedRunningTime="2025-12-10 19:21:31.539977377 +0000 UTC m=+1574.334825199" watchObservedRunningTime="2025-12-10 19:21:31.547822624 +0000 UTC m=+1574.342670426" Dec 10 19:21:34 crc kubenswrapper[4894]: I1210 19:21:34.094515 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:34 crc kubenswrapper[4894]: I1210 19:21:34.094808 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:35 crc kubenswrapper[4894]: I1210 19:21:35.159311 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-czlzc" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="registry-server" probeResult="failure" output=< Dec 10 19:21:35 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:21:35 crc kubenswrapper[4894]: > Dec 10 19:21:38 crc kubenswrapper[4894]: I1210 19:21:38.651965 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:38 crc kubenswrapper[4894]: I1210 19:21:38.652672 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:38 crc kubenswrapper[4894]: I1210 19:21:38.698570 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:39 crc kubenswrapper[4894]: I1210 19:21:39.670391 4894 generic.go:334] "Generic (PLEG): container finished" podID="919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" containerID="0ce2413aca8bf7581de470b73ffaf44bd9b6d1151c0665d86a225ee43d0b87c1" exitCode=0 Dec 10 19:21:39 crc kubenswrapper[4894]: I1210 19:21:39.670463 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" event={"ID":"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe","Type":"ContainerDied","Data":"0ce2413aca8bf7581de470b73ffaf44bd9b6d1151c0665d86a225ee43d0b87c1"} Dec 10 19:21:39 crc kubenswrapper[4894]: I1210 19:21:39.733538 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:39 crc kubenswrapper[4894]: I1210 19:21:39.789827 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8rsv"] Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.418339 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.514111 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-ssh-key\") pod \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.514695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkr7g\" (UniqueName: \"kubernetes.io/projected/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-kube-api-access-bkr7g\") pod \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.514750 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-repo-setup-combined-ca-bundle\") pod \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.514919 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-inventory\") pod \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\" (UID: \"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe\") " Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.521860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" (UID: "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.523372 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-kube-api-access-bkr7g" (OuterVolumeSpecName: "kube-api-access-bkr7g") pod "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" (UID: "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe"). InnerVolumeSpecName "kube-api-access-bkr7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.549636 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-inventory" (OuterVolumeSpecName: "inventory") pod "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" (UID: "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.560877 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" (UID: "919d6c2a-def8-4ef2-b1b0-9f65d61be6fe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.617803 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.617877 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.617890 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkr7g\" (UniqueName: \"kubernetes.io/projected/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-kube-api-access-bkr7g\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.617903 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/919d6c2a-def8-4ef2-b1b0-9f65d61be6fe-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.694771 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" event={"ID":"919d6c2a-def8-4ef2-b1b0-9f65d61be6fe","Type":"ContainerDied","Data":"e5dcb24e820754e1af506501afebe4bff7b0f6a5a1097ed818df672d66ee733b"} Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.694809 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lth47" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.694821 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5dcb24e820754e1af506501afebe4bff7b0f6a5a1097ed818df672d66ee733b" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.694972 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b8rsv" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="registry-server" containerID="cri-o://c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435" gracePeriod=2 Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.784034 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp"] Dec 10 19:21:41 crc kubenswrapper[4894]: E1210 19:21:41.784755 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.784792 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.785307 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="919d6c2a-def8-4ef2-b1b0-9f65d61be6fe" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.786526 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.793261 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.793366 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.793528 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.797552 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp"] Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.817815 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.925002 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.925372 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jlmc\" (UniqueName: \"kubernetes.io/projected/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-kube-api-access-8jlmc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:41 crc kubenswrapper[4894]: I1210 19:21:41.925496 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.027538 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.028574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.028606 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jlmc\" (UniqueName: \"kubernetes.io/projected/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-kube-api-access-8jlmc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.033025 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.033196 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.048389 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jlmc\" (UniqueName: \"kubernetes.io/projected/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-kube-api-access-8jlmc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-dwdzp\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.128001 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.162330 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.231934 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvjzp\" (UniqueName: \"kubernetes.io/projected/3d8ff699-e886-4443-b13d-257df5ff62f0-kube-api-access-wvjzp\") pod \"3d8ff699-e886-4443-b13d-257df5ff62f0\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.231997 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-catalog-content\") pod \"3d8ff699-e886-4443-b13d-257df5ff62f0\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.232017 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-utilities\") pod \"3d8ff699-e886-4443-b13d-257df5ff62f0\" (UID: \"3d8ff699-e886-4443-b13d-257df5ff62f0\") " Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.232914 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-utilities" (OuterVolumeSpecName: "utilities") pod "3d8ff699-e886-4443-b13d-257df5ff62f0" (UID: "3d8ff699-e886-4443-b13d-257df5ff62f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.235029 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d8ff699-e886-4443-b13d-257df5ff62f0-kube-api-access-wvjzp" (OuterVolumeSpecName: "kube-api-access-wvjzp") pod "3d8ff699-e886-4443-b13d-257df5ff62f0" (UID: "3d8ff699-e886-4443-b13d-257df5ff62f0"). InnerVolumeSpecName "kube-api-access-wvjzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.279952 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d8ff699-e886-4443-b13d-257df5ff62f0" (UID: "3d8ff699-e886-4443-b13d-257df5ff62f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.334663 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvjzp\" (UniqueName: \"kubernetes.io/projected/3d8ff699-e886-4443-b13d-257df5ff62f0-kube-api-access-wvjzp\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.334692 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.334705 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d8ff699-e886-4443-b13d-257df5ff62f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.709021 4894 generic.go:334] "Generic (PLEG): container finished" podID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerID="c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435" exitCode=0 Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.709084 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerDied","Data":"c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435"} Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.709119 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8rsv" event={"ID":"3d8ff699-e886-4443-b13d-257df5ff62f0","Type":"ContainerDied","Data":"bc77e3a1d3e6a07db24cc29e3afd5d2c297e85a248454937d44e367627033741"} Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.709145 4894 scope.go:117] "RemoveContainer" containerID="c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.709352 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8rsv" Dec 10 19:21:42 crc kubenswrapper[4894]: W1210 19:21:42.731447 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc05ca8d_4580_44b9_aad0_b23ce366f7ab.slice/crio-7b9c883a26ecf3847db3f9e762ed7de68be826ee93bfd14ec46d8f60a87cfbaf WatchSource:0}: Error finding container 7b9c883a26ecf3847db3f9e762ed7de68be826ee93bfd14ec46d8f60a87cfbaf: Status 404 returned error can't find the container with id 7b9c883a26ecf3847db3f9e762ed7de68be826ee93bfd14ec46d8f60a87cfbaf Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.734025 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp"] Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.740425 4894 scope.go:117] "RemoveContainer" containerID="5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.793261 4894 scope.go:117] "RemoveContainer" containerID="9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.853227 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8rsv"] Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.863679 4894 scope.go:117] "RemoveContainer" containerID="c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435" Dec 10 19:21:42 crc kubenswrapper[4894]: E1210 19:21:42.864361 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435\": container with ID starting with c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435 not found: ID does not exist" containerID="c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.864401 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435"} err="failed to get container status \"c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435\": rpc error: code = NotFound desc = could not find container \"c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435\": container with ID starting with c1b1cebf2f02f46433dd85a3903651c835e67a69b6c0efffb175c0e8e5287435 not found: ID does not exist" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.864428 4894 scope.go:117] "RemoveContainer" containerID="5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e" Dec 10 19:21:42 crc kubenswrapper[4894]: E1210 19:21:42.864816 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e\": container with ID starting with 5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e not found: ID does not exist" containerID="5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.864862 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e"} err="failed to get container status \"5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e\": rpc error: code = NotFound desc = could not find container \"5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e\": container with ID starting with 5984761121b029d79611ef317e7182b1f37199d5b6c25a7d9f5d842c0367039e not found: ID does not exist" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.864884 4894 scope.go:117] "RemoveContainer" containerID="9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.865021 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b8rsv"] Dec 10 19:21:42 crc kubenswrapper[4894]: E1210 19:21:42.865310 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f\": container with ID starting with 9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f not found: ID does not exist" containerID="9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f" Dec 10 19:21:42 crc kubenswrapper[4894]: I1210 19:21:42.865340 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f"} err="failed to get container status \"9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f\": rpc error: code = NotFound desc = could not find container \"9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f\": container with ID starting with 9bf3c0eb61493618432e316a8abb7cc6a2bb1398773def79a6cc689d7122f96f not found: ID does not exist" Dec 10 19:21:43 crc kubenswrapper[4894]: I1210 19:21:43.561375 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" path="/var/lib/kubelet/pods/3d8ff699-e886-4443-b13d-257df5ff62f0/volumes" Dec 10 19:21:43 crc kubenswrapper[4894]: I1210 19:21:43.720395 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" event={"ID":"fc05ca8d-4580-44b9-aad0-b23ce366f7ab","Type":"ContainerStarted","Data":"7b9c883a26ecf3847db3f9e762ed7de68be826ee93bfd14ec46d8f60a87cfbaf"} Dec 10 19:21:44 crc kubenswrapper[4894]: I1210 19:21:44.194375 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:44 crc kubenswrapper[4894]: I1210 19:21:44.267048 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:44 crc kubenswrapper[4894]: I1210 19:21:44.740061 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" event={"ID":"fc05ca8d-4580-44b9-aad0-b23ce366f7ab","Type":"ContainerStarted","Data":"4e5e068be77c37dd07c4cef4d5072a1d888998edf7401e7d0fa863e0ebfdf672"} Dec 10 19:21:44 crc kubenswrapper[4894]: I1210 19:21:44.766792 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" podStartSLOduration=3.058572722 podStartE2EDuration="3.766766764s" podCreationTimestamp="2025-12-10 19:21:41 +0000 UTC" firstStartedPulling="2025-12-10 19:21:42.740452193 +0000 UTC m=+1585.535299985" lastFinishedPulling="2025-12-10 19:21:43.448646235 +0000 UTC m=+1586.243494027" observedRunningTime="2025-12-10 19:21:44.755259405 +0000 UTC m=+1587.550107197" watchObservedRunningTime="2025-12-10 19:21:44.766766764 +0000 UTC m=+1587.561614566" Dec 10 19:21:45 crc kubenswrapper[4894]: I1210 19:21:45.382733 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-czlzc"] Dec 10 19:21:45 crc kubenswrapper[4894]: I1210 19:21:45.750209 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-czlzc" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="registry-server" containerID="cri-o://13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b" gracePeriod=2 Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.300766 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.427062 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-utilities\") pod \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.427193 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-catalog-content\") pod \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.427338 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjs4c\" (UniqueName: \"kubernetes.io/projected/1d18d063-dbd5-41bc-88a5-db817a1b75bb-kube-api-access-rjs4c\") pod \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\" (UID: \"1d18d063-dbd5-41bc-88a5-db817a1b75bb\") " Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.427897 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-utilities" (OuterVolumeSpecName: "utilities") pod "1d18d063-dbd5-41bc-88a5-db817a1b75bb" (UID: "1d18d063-dbd5-41bc-88a5-db817a1b75bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.428109 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.435223 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d18d063-dbd5-41bc-88a5-db817a1b75bb-kube-api-access-rjs4c" (OuterVolumeSpecName: "kube-api-access-rjs4c") pod "1d18d063-dbd5-41bc-88a5-db817a1b75bb" (UID: "1d18d063-dbd5-41bc-88a5-db817a1b75bb"). InnerVolumeSpecName "kube-api-access-rjs4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.498026 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d18d063-dbd5-41bc-88a5-db817a1b75bb" (UID: "1d18d063-dbd5-41bc-88a5-db817a1b75bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.529598 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d18d063-dbd5-41bc-88a5-db817a1b75bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.529630 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjs4c\" (UniqueName: \"kubernetes.io/projected/1d18d063-dbd5-41bc-88a5-db817a1b75bb-kube-api-access-rjs4c\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.763833 4894 generic.go:334] "Generic (PLEG): container finished" podID="fc05ca8d-4580-44b9-aad0-b23ce366f7ab" containerID="4e5e068be77c37dd07c4cef4d5072a1d888998edf7401e7d0fa863e0ebfdf672" exitCode=0 Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.763955 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" event={"ID":"fc05ca8d-4580-44b9-aad0-b23ce366f7ab","Type":"ContainerDied","Data":"4e5e068be77c37dd07c4cef4d5072a1d888998edf7401e7d0fa863e0ebfdf672"} Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.770140 4894 generic.go:334] "Generic (PLEG): container finished" podID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerID="13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b" exitCode=0 Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.770213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerDied","Data":"13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b"} Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.770257 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czlzc" event={"ID":"1d18d063-dbd5-41bc-88a5-db817a1b75bb","Type":"ContainerDied","Data":"205ea045d87212c96217901987a693a578c129ed38ad4735852e7c8e77304370"} Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.770287 4894 scope.go:117] "RemoveContainer" containerID="13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.770518 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czlzc" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.823818 4894 scope.go:117] "RemoveContainer" containerID="33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.864999 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-czlzc"] Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.867059 4894 scope.go:117] "RemoveContainer" containerID="bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.879621 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-czlzc"] Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.940825 4894 scope.go:117] "RemoveContainer" containerID="13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b" Dec 10 19:21:46 crc kubenswrapper[4894]: E1210 19:21:46.941225 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b\": container with ID starting with 13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b not found: ID does not exist" containerID="13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.941252 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b"} err="failed to get container status \"13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b\": rpc error: code = NotFound desc = could not find container \"13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b\": container with ID starting with 13a82e83a59cac238d3a1199d88ff9cb00a4e6de1cf0971d425b184954e60d0b not found: ID does not exist" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.941272 4894 scope.go:117] "RemoveContainer" containerID="33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2" Dec 10 19:21:46 crc kubenswrapper[4894]: E1210 19:21:46.941748 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2\": container with ID starting with 33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2 not found: ID does not exist" containerID="33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.941771 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2"} err="failed to get container status \"33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2\": rpc error: code = NotFound desc = could not find container \"33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2\": container with ID starting with 33d66788882e6565f241b1b9fe5d94cb986fbb49b87d9b1d6b02313f58b0e9f2 not found: ID does not exist" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.941783 4894 scope.go:117] "RemoveContainer" containerID="bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f" Dec 10 19:21:46 crc kubenswrapper[4894]: E1210 19:21:46.942035 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f\": container with ID starting with bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f not found: ID does not exist" containerID="bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f" Dec 10 19:21:46 crc kubenswrapper[4894]: I1210 19:21:46.942052 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f"} err="failed to get container status \"bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f\": rpc error: code = NotFound desc = could not find container \"bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f\": container with ID starting with bdc3706d9f9f48fcfeecf51f61e9b0cb48884d82b704ccb743299378ec07512f not found: ID does not exist" Dec 10 19:21:47 crc kubenswrapper[4894]: E1210 19:21:47.007944 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d18d063_dbd5_41bc_88a5_db817a1b75bb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d18d063_dbd5_41bc_88a5_db817a1b75bb.slice/crio-205ea045d87212c96217901987a693a578c129ed38ad4735852e7c8e77304370\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:47 crc kubenswrapper[4894]: I1210 19:21:47.558488 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" path="/var/lib/kubelet/pods/1d18d063-dbd5-41bc-88a5-db817a1b75bb/volumes" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.373356 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.398232 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-inventory\") pod \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.398306 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jlmc\" (UniqueName: \"kubernetes.io/projected/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-kube-api-access-8jlmc\") pod \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.398444 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-ssh-key\") pod \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\" (UID: \"fc05ca8d-4580-44b9-aad0-b23ce366f7ab\") " Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.414306 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-kube-api-access-8jlmc" (OuterVolumeSpecName: "kube-api-access-8jlmc") pod "fc05ca8d-4580-44b9-aad0-b23ce366f7ab" (UID: "fc05ca8d-4580-44b9-aad0-b23ce366f7ab"). InnerVolumeSpecName "kube-api-access-8jlmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.442779 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-inventory" (OuterVolumeSpecName: "inventory") pod "fc05ca8d-4580-44b9-aad0-b23ce366f7ab" (UID: "fc05ca8d-4580-44b9-aad0-b23ce366f7ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.443117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc05ca8d-4580-44b9-aad0-b23ce366f7ab" (UID: "fc05ca8d-4580-44b9-aad0-b23ce366f7ab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.501038 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.501087 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.501099 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jlmc\" (UniqueName: \"kubernetes.io/projected/fc05ca8d-4580-44b9-aad0-b23ce366f7ab-kube-api-access-8jlmc\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.801002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" event={"ID":"fc05ca8d-4580-44b9-aad0-b23ce366f7ab","Type":"ContainerDied","Data":"7b9c883a26ecf3847db3f9e762ed7de68be826ee93bfd14ec46d8f60a87cfbaf"} Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.801045 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b9c883a26ecf3847db3f9e762ed7de68be826ee93bfd14ec46d8f60a87cfbaf" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.801077 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-dwdzp" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.930962 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46"] Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953172 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="registry-server" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953211 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="registry-server" Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953242 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="extract-content" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953249 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="extract-content" Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953267 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="extract-utilities" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953273 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="extract-utilities" Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953315 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="extract-utilities" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953322 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="extract-utilities" Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953339 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc05ca8d-4580-44b9-aad0-b23ce366f7ab" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953346 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc05ca8d-4580-44b9-aad0-b23ce366f7ab" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953361 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="extract-content" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953366 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="extract-content" Dec 10 19:21:48 crc kubenswrapper[4894]: E1210 19:21:48.953384 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="registry-server" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.953389 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="registry-server" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.954036 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d8ff699-e886-4443-b13d-257df5ff62f0" containerName="registry-server" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.954076 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d18d063-dbd5-41bc-88a5-db817a1b75bb" containerName="registry-server" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.954092 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc05ca8d-4580-44b9-aad0-b23ce366f7ab" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.955051 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.978581 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.978818 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.978978 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:21:48 crc kubenswrapper[4894]: I1210 19:21:48.979147 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.024300 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.024379 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.024964 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c8l7\" (UniqueName: \"kubernetes.io/projected/b1c310ef-2846-4536-9978-2cfd1c9a606c-kube-api-access-5c8l7\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.025010 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.064598 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46"] Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.127429 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c8l7\" (UniqueName: \"kubernetes.io/projected/b1c310ef-2846-4536-9978-2cfd1c9a606c-kube-api-access-5c8l7\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.127673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.127725 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.127775 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.131731 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.134308 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.141532 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.144451 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c8l7\" (UniqueName: \"kubernetes.io/projected/b1c310ef-2846-4536-9978-2cfd1c9a606c-kube-api-access-5c8l7\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.333665 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.584319 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.584724 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:21:49 crc kubenswrapper[4894]: I1210 19:21:49.923421 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46"] Dec 10 19:21:50 crc kubenswrapper[4894]: I1210 19:21:50.821193 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" event={"ID":"b1c310ef-2846-4536-9978-2cfd1c9a606c","Type":"ContainerStarted","Data":"d5a066b5fa83dfccbadd51de3c50662fc26fc78eab43999e197bb89859e278ab"} Dec 10 19:21:51 crc kubenswrapper[4894]: I1210 19:21:51.836777 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" event={"ID":"b1c310ef-2846-4536-9978-2cfd1c9a606c","Type":"ContainerStarted","Data":"385ff8da671bf132e631d457bdeadadb6a4d5469c004362dee644f806de38fbb"} Dec 10 19:21:51 crc kubenswrapper[4894]: I1210 19:21:51.878835 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" podStartSLOduration=3.2355188 podStartE2EDuration="3.87881004s" podCreationTimestamp="2025-12-10 19:21:48 +0000 UTC" firstStartedPulling="2025-12-10 19:21:49.935667497 +0000 UTC m=+1592.730515289" lastFinishedPulling="2025-12-10 19:21:50.578958727 +0000 UTC m=+1593.373806529" observedRunningTime="2025-12-10 19:21:51.871468818 +0000 UTC m=+1594.666316650" watchObservedRunningTime="2025-12-10 19:21:51.87881004 +0000 UTC m=+1594.673657842" Dec 10 19:22:19 crc kubenswrapper[4894]: I1210 19:22:19.577079 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:22:19 crc kubenswrapper[4894]: I1210 19:22:19.577659 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:22:19 crc kubenswrapper[4894]: I1210 19:22:19.577701 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:22:19 crc kubenswrapper[4894]: I1210 19:22:19.578554 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:22:19 crc kubenswrapper[4894]: I1210 19:22:19.578609 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" gracePeriod=600 Dec 10 19:22:19 crc kubenswrapper[4894]: E1210 19:22:19.700053 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:22:19 crc kubenswrapper[4894]: I1210 19:22:19.991801 4894 scope.go:117] "RemoveContainer" containerID="0c743f4c3522c30eaf964cce3270510dd52fdcb74163f7f9e3730046b5c6af12" Dec 10 19:22:20 crc kubenswrapper[4894]: I1210 19:22:20.022663 4894 scope.go:117] "RemoveContainer" containerID="34c65a66b3db9b6eac6c97117490d462700c7f0cea9df74fa545c757a24f8484" Dec 10 19:22:20 crc kubenswrapper[4894]: I1210 19:22:20.069827 4894 scope.go:117] "RemoveContainer" containerID="0cb0a7113b5b12e095db65306bf061698d8600b4438f43afad3e4c4751eaf9a2" Dec 10 19:22:20 crc kubenswrapper[4894]: I1210 19:22:20.171727 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" exitCode=0 Dec 10 19:22:20 crc kubenswrapper[4894]: I1210 19:22:20.171916 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4"} Dec 10 19:22:20 crc kubenswrapper[4894]: I1210 19:22:20.171996 4894 scope.go:117] "RemoveContainer" containerID="5c62314c4a97570eb40f37c7f81fec8ac3dc6ad9954e24945138dbb31ee1ce97" Dec 10 19:22:20 crc kubenswrapper[4894]: I1210 19:22:20.173241 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:22:20 crc kubenswrapper[4894]: E1210 19:22:20.174098 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:22:33 crc kubenswrapper[4894]: I1210 19:22:33.542294 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:22:33 crc kubenswrapper[4894]: E1210 19:22:33.547707 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:22:45 crc kubenswrapper[4894]: I1210 19:22:45.541752 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:22:45 crc kubenswrapper[4894]: E1210 19:22:45.543064 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:22:59 crc kubenswrapper[4894]: I1210 19:22:59.541682 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:22:59 crc kubenswrapper[4894]: E1210 19:22:59.542650 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:23:10 crc kubenswrapper[4894]: I1210 19:23:10.541556 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:23:10 crc kubenswrapper[4894]: E1210 19:23:10.542750 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:23:20 crc kubenswrapper[4894]: I1210 19:23:20.219390 4894 scope.go:117] "RemoveContainer" containerID="428fe02775bef18d060f3c3ecbbd4cbb20b6c059bfa694e2f62243ed2fbb2e90" Dec 10 19:23:20 crc kubenswrapper[4894]: I1210 19:23:20.259855 4894 scope.go:117] "RemoveContainer" containerID="3a3033834bef69f8782e1cff647ddcbc0d44dff33e2bb1c8d8f91898c281d0c9" Dec 10 19:23:24 crc kubenswrapper[4894]: I1210 19:23:24.540990 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:23:24 crc kubenswrapper[4894]: E1210 19:23:24.542032 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:23:36 crc kubenswrapper[4894]: I1210 19:23:36.540669 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:23:36 crc kubenswrapper[4894]: E1210 19:23:36.541565 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:23:49 crc kubenswrapper[4894]: I1210 19:23:49.542567 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:23:49 crc kubenswrapper[4894]: E1210 19:23:49.543486 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:24:01 crc kubenswrapper[4894]: I1210 19:24:01.542104 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:24:01 crc kubenswrapper[4894]: E1210 19:24:01.543246 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:24:16 crc kubenswrapper[4894]: I1210 19:24:16.541250 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:24:16 crc kubenswrapper[4894]: E1210 19:24:16.543314 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:24:20 crc kubenswrapper[4894]: I1210 19:24:20.424988 4894 scope.go:117] "RemoveContainer" containerID="fcd772eee6a7ebb5c88934d022ebd48ecddd199bbee9c9360156d698935d92e1" Dec 10 19:24:20 crc kubenswrapper[4894]: I1210 19:24:20.465649 4894 scope.go:117] "RemoveContainer" containerID="7ae3372213b38e954187c48bc245885342733013650237dea1649cadbf8237a4" Dec 10 19:24:20 crc kubenswrapper[4894]: I1210 19:24:20.492701 4894 scope.go:117] "RemoveContainer" containerID="885c37f3498f8bbf498badf4e3529e8cc0959aaa9cd8b6f56b38a0a27300bf4e" Dec 10 19:24:20 crc kubenswrapper[4894]: I1210 19:24:20.514454 4894 scope.go:117] "RemoveContainer" containerID="bfc993891b2abfaea1dad50a4536709dd87497086cd7b8728c9859894d362fec" Dec 10 19:24:20 crc kubenswrapper[4894]: I1210 19:24:20.539765 4894 scope.go:117] "RemoveContainer" containerID="bf26864ee39cae57e5fcab67a75c02b0a0f9ecff7ffae67b6557c2610c66368d" Dec 10 19:24:20 crc kubenswrapper[4894]: I1210 19:24:20.568779 4894 scope.go:117] "RemoveContainer" containerID="2a56d533fda516fbd7e25da758f2323d2e62776ec8b4ad563cdd351c18f78582" Dec 10 19:24:31 crc kubenswrapper[4894]: I1210 19:24:31.540833 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:24:31 crc kubenswrapper[4894]: E1210 19:24:31.541684 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:24:44 crc kubenswrapper[4894]: I1210 19:24:44.541335 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:24:44 crc kubenswrapper[4894]: E1210 19:24:44.542041 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:24:56 crc kubenswrapper[4894]: I1210 19:24:56.540944 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:24:56 crc kubenswrapper[4894]: E1210 19:24:56.541671 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:25:08 crc kubenswrapper[4894]: I1210 19:25:08.488833 4894 generic.go:334] "Generic (PLEG): container finished" podID="b1c310ef-2846-4536-9978-2cfd1c9a606c" containerID="385ff8da671bf132e631d457bdeadadb6a4d5469c004362dee644f806de38fbb" exitCode=0 Dec 10 19:25:08 crc kubenswrapper[4894]: I1210 19:25:08.488987 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" event={"ID":"b1c310ef-2846-4536-9978-2cfd1c9a606c","Type":"ContainerDied","Data":"385ff8da671bf132e631d457bdeadadb6a4d5469c004362dee644f806de38fbb"} Dec 10 19:25:08 crc kubenswrapper[4894]: I1210 19:25:08.540868 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:25:08 crc kubenswrapper[4894]: E1210 19:25:08.541283 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.030103 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.140831 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-ssh-key\") pod \"b1c310ef-2846-4536-9978-2cfd1c9a606c\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.140982 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-inventory\") pod \"b1c310ef-2846-4536-9978-2cfd1c9a606c\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.141127 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c8l7\" (UniqueName: \"kubernetes.io/projected/b1c310ef-2846-4536-9978-2cfd1c9a606c-kube-api-access-5c8l7\") pod \"b1c310ef-2846-4536-9978-2cfd1c9a606c\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.141204 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-bootstrap-combined-ca-bundle\") pod \"b1c310ef-2846-4536-9978-2cfd1c9a606c\" (UID: \"b1c310ef-2846-4536-9978-2cfd1c9a606c\") " Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.146337 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b1c310ef-2846-4536-9978-2cfd1c9a606c" (UID: "b1c310ef-2846-4536-9978-2cfd1c9a606c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.146455 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c310ef-2846-4536-9978-2cfd1c9a606c-kube-api-access-5c8l7" (OuterVolumeSpecName: "kube-api-access-5c8l7") pod "b1c310ef-2846-4536-9978-2cfd1c9a606c" (UID: "b1c310ef-2846-4536-9978-2cfd1c9a606c"). InnerVolumeSpecName "kube-api-access-5c8l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.173384 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1c310ef-2846-4536-9978-2cfd1c9a606c" (UID: "b1c310ef-2846-4536-9978-2cfd1c9a606c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.175636 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-inventory" (OuterVolumeSpecName: "inventory") pod "b1c310ef-2846-4536-9978-2cfd1c9a606c" (UID: "b1c310ef-2846-4536-9978-2cfd1c9a606c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.243560 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c8l7\" (UniqueName: \"kubernetes.io/projected/b1c310ef-2846-4536-9978-2cfd1c9a606c-kube-api-access-5c8l7\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.243591 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.243600 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.243609 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1c310ef-2846-4536-9978-2cfd1c9a606c-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.513157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" event={"ID":"b1c310ef-2846-4536-9978-2cfd1c9a606c","Type":"ContainerDied","Data":"d5a066b5fa83dfccbadd51de3c50662fc26fc78eab43999e197bb89859e278ab"} Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.513203 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5a066b5fa83dfccbadd51de3c50662fc26fc78eab43999e197bb89859e278ab" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.513202 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.615246 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt"] Dec 10 19:25:10 crc kubenswrapper[4894]: E1210 19:25:10.615712 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c310ef-2846-4536-9978-2cfd1c9a606c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.615730 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c310ef-2846-4536-9978-2cfd1c9a606c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.615929 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c310ef-2846-4536-9978-2cfd1c9a606c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.616679 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.619513 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.620172 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.620669 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.620670 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.630927 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt"] Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.754338 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.754385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.754470 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtmt7\" (UniqueName: \"kubernetes.io/projected/25aa478e-2e66-4ba8-9428-17ad67e64df0-kube-api-access-mtmt7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.858019 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.859425 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.859620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtmt7\" (UniqueName: \"kubernetes.io/projected/25aa478e-2e66-4ba8-9428-17ad67e64df0-kube-api-access-mtmt7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.865035 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.866338 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.879744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtmt7\" (UniqueName: \"kubernetes.io/projected/25aa478e-2e66-4ba8-9428-17ad67e64df0-kube-api-access-mtmt7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:10 crc kubenswrapper[4894]: I1210 19:25:10.979242 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:25:11 crc kubenswrapper[4894]: I1210 19:25:11.531604 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt"] Dec 10 19:25:11 crc kubenswrapper[4894]: W1210 19:25:11.538026 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25aa478e_2e66_4ba8_9428_17ad67e64df0.slice/crio-d64a2ea6ff9209be8a1d7b5221d2803ee2d9569ee98b60a5a80d69158262bacb WatchSource:0}: Error finding container d64a2ea6ff9209be8a1d7b5221d2803ee2d9569ee98b60a5a80d69158262bacb: Status 404 returned error can't find the container with id d64a2ea6ff9209be8a1d7b5221d2803ee2d9569ee98b60a5a80d69158262bacb Dec 10 19:25:11 crc kubenswrapper[4894]: I1210 19:25:11.541341 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:25:12 crc kubenswrapper[4894]: I1210 19:25:12.531755 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" event={"ID":"25aa478e-2e66-4ba8-9428-17ad67e64df0","Type":"ContainerStarted","Data":"e0e74e52177aad84fb60a4fabaea4913b143a729da3aedeec14a5e0905202817"} Dec 10 19:25:12 crc kubenswrapper[4894]: I1210 19:25:12.532063 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" event={"ID":"25aa478e-2e66-4ba8-9428-17ad67e64df0","Type":"ContainerStarted","Data":"d64a2ea6ff9209be8a1d7b5221d2803ee2d9569ee98b60a5a80d69158262bacb"} Dec 10 19:25:12 crc kubenswrapper[4894]: I1210 19:25:12.550150 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" podStartSLOduration=2.110460694 podStartE2EDuration="2.550128717s" podCreationTimestamp="2025-12-10 19:25:10 +0000 UTC" firstStartedPulling="2025-12-10 19:25:11.541051391 +0000 UTC m=+1794.335899183" lastFinishedPulling="2025-12-10 19:25:11.980719394 +0000 UTC m=+1794.775567206" observedRunningTime="2025-12-10 19:25:12.545477571 +0000 UTC m=+1795.340325363" watchObservedRunningTime="2025-12-10 19:25:12.550128717 +0000 UTC m=+1795.344976509" Dec 10 19:25:21 crc kubenswrapper[4894]: I1210 19:25:21.541658 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:25:21 crc kubenswrapper[4894]: E1210 19:25:21.543495 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:25:36 crc kubenswrapper[4894]: I1210 19:25:36.541831 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:25:36 crc kubenswrapper[4894]: E1210 19:25:36.542891 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:25:50 crc kubenswrapper[4894]: I1210 19:25:50.541241 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:25:50 crc kubenswrapper[4894]: E1210 19:25:50.542108 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:25:53 crc kubenswrapper[4894]: I1210 19:25:53.070176 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9723-account-create-update-4b25k"] Dec 10 19:25:53 crc kubenswrapper[4894]: I1210 19:25:53.090081 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9723-account-create-update-4b25k"] Dec 10 19:25:53 crc kubenswrapper[4894]: I1210 19:25:53.561995 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3578f846-54cc-4fa8-8d9e-3dacbbf273db" path="/var/lib/kubelet/pods/3578f846-54cc-4fa8-8d9e-3dacbbf273db/volumes" Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.039753 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lsxb5"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.053911 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-d80b-account-create-update-9w4rh"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.065158 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lsxb5"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.073996 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-pgjdk"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.082523 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-pgjdk"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.109048 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-d80b-account-create-update-9w4rh"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.122404 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4c8b-account-create-update-wdpnl"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.134074 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-vq968"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.143886 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4c8b-account-create-update-wdpnl"] Dec 10 19:25:54 crc kubenswrapper[4894]: I1210 19:25:54.154574 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-vq968"] Dec 10 19:25:55 crc kubenswrapper[4894]: I1210 19:25:55.560601 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12383739-aa0a-4040-8d8d-f28fe4f155d0" path="/var/lib/kubelet/pods/12383739-aa0a-4040-8d8d-f28fe4f155d0/volumes" Dec 10 19:25:55 crc kubenswrapper[4894]: I1210 19:25:55.561637 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb80a4c-f422-4623-841b-ab19ea29f567" path="/var/lib/kubelet/pods/1eb80a4c-f422-4623-841b-ab19ea29f567/volumes" Dec 10 19:25:55 crc kubenswrapper[4894]: I1210 19:25:55.562451 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538ed591-7bb5-4109-942e-6a5b83171db4" path="/var/lib/kubelet/pods/538ed591-7bb5-4109-942e-6a5b83171db4/volumes" Dec 10 19:25:55 crc kubenswrapper[4894]: I1210 19:25:55.563249 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85391f2b-d55c-4abf-a95f-562c9be55a78" path="/var/lib/kubelet/pods/85391f2b-d55c-4abf-a95f-562c9be55a78/volumes" Dec 10 19:25:55 crc kubenswrapper[4894]: I1210 19:25:55.564937 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="950c46f1-26b7-4213-9db5-e49f5cd4ae31" path="/var/lib/kubelet/pods/950c46f1-26b7-4213-9db5-e49f5cd4ae31/volumes" Dec 10 19:26:03 crc kubenswrapper[4894]: I1210 19:26:03.541917 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:26:03 crc kubenswrapper[4894]: E1210 19:26:03.542676 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.044695 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-sq9hc"] Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.054089 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-58cb-account-create-update-n2g6b"] Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.063321 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-58cb-account-create-update-n2g6b"] Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.072556 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-sq9hc"] Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.554773 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:26:17 crc kubenswrapper[4894]: E1210 19:26:17.555559 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.557148 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224c4995-4a06-4e42-bd64-39794296a221" path="/var/lib/kubelet/pods/224c4995-4a06-4e42-bd64-39794296a221/volumes" Dec 10 19:26:17 crc kubenswrapper[4894]: I1210 19:26:17.557860 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd762222-8af2-4ce2-91a4-eb230aacda33" path="/var/lib/kubelet/pods/fd762222-8af2-4ce2-91a4-eb230aacda33/volumes" Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.068214 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-s5x8c"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.083758 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2ltzr"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.094682 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-hc6p7"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.103764 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-s5x8c"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.111768 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2ltzr"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.120079 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-4a1c-account-create-update-lrw4k"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.129114 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-ade3-account-create-update-cbv5v"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.139888 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8c45-account-create-update-6d449"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.151096 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-hc6p7"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.160793 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8c45-account-create-update-6d449"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.171295 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-4a1c-account-create-update-lrw4k"] Dec 10 19:26:18 crc kubenswrapper[4894]: I1210 19:26:18.181644 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-ade3-account-create-update-cbv5v"] Dec 10 19:26:19 crc kubenswrapper[4894]: I1210 19:26:19.551805 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f4e887-2f35-4567-b9be-931ff8e6b52c" path="/var/lib/kubelet/pods/45f4e887-2f35-4567-b9be-931ff8e6b52c/volumes" Dec 10 19:26:19 crc kubenswrapper[4894]: I1210 19:26:19.553043 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ed755d-307b-43d6-b84f-24c5f7284f24" path="/var/lib/kubelet/pods/86ed755d-307b-43d6-b84f-24c5f7284f24/volumes" Dec 10 19:26:19 crc kubenswrapper[4894]: I1210 19:26:19.553568 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf58f8ea-b763-4a18-9697-865f5741e5c7" path="/var/lib/kubelet/pods/cf58f8ea-b763-4a18-9697-865f5741e5c7/volumes" Dec 10 19:26:19 crc kubenswrapper[4894]: I1210 19:26:19.554193 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d500410c-9685-4529-b187-750ea33d7b99" path="/var/lib/kubelet/pods/d500410c-9685-4529-b187-750ea33d7b99/volumes" Dec 10 19:26:19 crc kubenswrapper[4894]: I1210 19:26:19.555333 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec67bfcd-9767-4e89-9683-0fe681539e23" path="/var/lib/kubelet/pods/ec67bfcd-9767-4e89-9683-0fe681539e23/volumes" Dec 10 19:26:19 crc kubenswrapper[4894]: I1210 19:26:19.555899 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f60159f7-610c-4277-9f46-7551f1ec247d" path="/var/lib/kubelet/pods/f60159f7-610c-4277-9f46-7551f1ec247d/volumes" Dec 10 19:26:20 crc kubenswrapper[4894]: I1210 19:26:20.771342 4894 scope.go:117] "RemoveContainer" containerID="620148e89fa9a80437e93e53da7fa560759cccbed1c2d01484633e1ace3ba2af" Dec 10 19:26:20 crc kubenswrapper[4894]: I1210 19:26:20.793644 4894 scope.go:117] "RemoveContainer" containerID="a6d372f049bdc761d0128659b5797103055988677c75f7f760bf046e5bee592c" Dec 10 19:26:20 crc kubenswrapper[4894]: I1210 19:26:20.851005 4894 scope.go:117] "RemoveContainer" containerID="2c23663e860f29555de487eb4deecf1f0e5773a2613775ee2661cc4a4be06f01" Dec 10 19:26:20 crc kubenswrapper[4894]: I1210 19:26:20.904636 4894 scope.go:117] "RemoveContainer" containerID="6f705d1f9870dc89ac0338837a84a52e179f9221451404bc6d3954ce938a56d0" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.031205 4894 scope.go:117] "RemoveContainer" containerID="7559549a1f572e5280f008033089971297e58264d079106ec602858ee9371b64" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.062627 4894 scope.go:117] "RemoveContainer" containerID="7532b830b73e3d7e460d3a094e3729ab502b06106dcb41cf3785962718d445f5" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.108862 4894 scope.go:117] "RemoveContainer" containerID="4fa854ec517d77864924b4341c079a285b769a523ab672b091a0676d8391da5a" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.130771 4894 scope.go:117] "RemoveContainer" containerID="971eb7e85dadeaca18768502288fa10f035cfb8796decafd3a4e66520388469f" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.150182 4894 scope.go:117] "RemoveContainer" containerID="6c6d56d2ec1fff10ae2fd378f0568fba0a7735260512f9d8fde45cfe141529a3" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.171199 4894 scope.go:117] "RemoveContainer" containerID="a4a8060b1c0e3b910adcc3bdab022f87796f5d1a9ae6916e69bac93e36a0c3d2" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.189413 4894 scope.go:117] "RemoveContainer" containerID="0273188ce1e0f0d6ebac343ae54ee9b66a251566c4820339e1c5badc6f6e4664" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.206715 4894 scope.go:117] "RemoveContainer" containerID="0acc7bbf26de8adf07f305399736ee77810e3f99a01d6faf2e0bc7426c5c2841" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.226321 4894 scope.go:117] "RemoveContainer" containerID="cf21f959697e074ec576f1d07143a2c73ede2db9611fc61d3cace6b9b0b82e92" Dec 10 19:26:21 crc kubenswrapper[4894]: I1210 19:26:21.244866 4894 scope.go:117] "RemoveContainer" containerID="6037d10d68cd85c0d7ba2af1c208575a5d2c5df33c6bbeccd2fb1e43cec12cf5" Dec 10 19:26:22 crc kubenswrapper[4894]: I1210 19:26:22.032301 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-pk77t"] Dec 10 19:26:22 crc kubenswrapper[4894]: I1210 19:26:22.047224 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-kxf26"] Dec 10 19:26:22 crc kubenswrapper[4894]: I1210 19:26:22.058950 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-pk77t"] Dec 10 19:26:22 crc kubenswrapper[4894]: I1210 19:26:22.071175 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-kxf26"] Dec 10 19:26:23 crc kubenswrapper[4894]: I1210 19:26:23.557945 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b4489ab-f1cb-4296-b5f7-6168189d3be0" path="/var/lib/kubelet/pods/0b4489ab-f1cb-4296-b5f7-6168189d3be0/volumes" Dec 10 19:26:23 crc kubenswrapper[4894]: I1210 19:26:23.559132 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f1bbcc0-1d98-452d-9b61-9caf78395cb4" path="/var/lib/kubelet/pods/8f1bbcc0-1d98-452d-9b61-9caf78395cb4/volumes" Dec 10 19:26:29 crc kubenswrapper[4894]: I1210 19:26:29.540713 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:26:29 crc kubenswrapper[4894]: E1210 19:26:29.541339 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:26:41 crc kubenswrapper[4894]: I1210 19:26:41.542554 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:26:41 crc kubenswrapper[4894]: E1210 19:26:41.544202 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:26:52 crc kubenswrapper[4894]: I1210 19:26:52.542461 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:26:52 crc kubenswrapper[4894]: E1210 19:26:52.543829 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:26:53 crc kubenswrapper[4894]: I1210 19:26:53.060919 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jk98n"] Dec 10 19:26:53 crc kubenswrapper[4894]: I1210 19:26:53.072958 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jk98n"] Dec 10 19:26:53 crc kubenswrapper[4894]: I1210 19:26:53.552448 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a158cd9-dd00-4d21-bc0d-cf9f84aca910" path="/var/lib/kubelet/pods/2a158cd9-dd00-4d21-bc0d-cf9f84aca910/volumes" Dec 10 19:27:05 crc kubenswrapper[4894]: I1210 19:27:05.866820 4894 generic.go:334] "Generic (PLEG): container finished" podID="25aa478e-2e66-4ba8-9428-17ad67e64df0" containerID="e0e74e52177aad84fb60a4fabaea4913b143a729da3aedeec14a5e0905202817" exitCode=0 Dec 10 19:27:05 crc kubenswrapper[4894]: I1210 19:27:05.867390 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" event={"ID":"25aa478e-2e66-4ba8-9428-17ad67e64df0","Type":"ContainerDied","Data":"e0e74e52177aad84fb60a4fabaea4913b143a729da3aedeec14a5e0905202817"} Dec 10 19:27:06 crc kubenswrapper[4894]: I1210 19:27:06.541891 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:27:06 crc kubenswrapper[4894]: E1210 19:27:06.542470 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.048795 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4gmfn"] Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.062677 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4gmfn"] Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.452795 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.558586 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a781080-8482-4915-9d13-b88d1785f7f3" path="/var/lib/kubelet/pods/7a781080-8482-4915-9d13-b88d1785f7f3/volumes" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.573053 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-ssh-key\") pod \"25aa478e-2e66-4ba8-9428-17ad67e64df0\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.573133 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-inventory\") pod \"25aa478e-2e66-4ba8-9428-17ad67e64df0\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.573338 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtmt7\" (UniqueName: \"kubernetes.io/projected/25aa478e-2e66-4ba8-9428-17ad67e64df0-kube-api-access-mtmt7\") pod \"25aa478e-2e66-4ba8-9428-17ad67e64df0\" (UID: \"25aa478e-2e66-4ba8-9428-17ad67e64df0\") " Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.582972 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25aa478e-2e66-4ba8-9428-17ad67e64df0-kube-api-access-mtmt7" (OuterVolumeSpecName: "kube-api-access-mtmt7") pod "25aa478e-2e66-4ba8-9428-17ad67e64df0" (UID: "25aa478e-2e66-4ba8-9428-17ad67e64df0"). InnerVolumeSpecName "kube-api-access-mtmt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.610136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-inventory" (OuterVolumeSpecName: "inventory") pod "25aa478e-2e66-4ba8-9428-17ad67e64df0" (UID: "25aa478e-2e66-4ba8-9428-17ad67e64df0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.611149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "25aa478e-2e66-4ba8-9428-17ad67e64df0" (UID: "25aa478e-2e66-4ba8-9428-17ad67e64df0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.676930 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.676969 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtmt7\" (UniqueName: \"kubernetes.io/projected/25aa478e-2e66-4ba8-9428-17ad67e64df0-kube-api-access-mtmt7\") on node \"crc\" DevicePath \"\"" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.676983 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25aa478e-2e66-4ba8-9428-17ad67e64df0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.891533 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" event={"ID":"25aa478e-2e66-4ba8-9428-17ad67e64df0","Type":"ContainerDied","Data":"d64a2ea6ff9209be8a1d7b5221d2803ee2d9569ee98b60a5a80d69158262bacb"} Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.891599 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d64a2ea6ff9209be8a1d7b5221d2803ee2d9569ee98b60a5a80d69158262bacb" Dec 10 19:27:07 crc kubenswrapper[4894]: I1210 19:27:07.891616 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.001442 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr"] Dec 10 19:27:08 crc kubenswrapper[4894]: E1210 19:27:08.001946 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25aa478e-2e66-4ba8-9428-17ad67e64df0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.001968 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="25aa478e-2e66-4ba8-9428-17ad67e64df0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.002223 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="25aa478e-2e66-4ba8-9428-17ad67e64df0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.003073 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.008167 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.008382 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.009180 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.014512 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.025249 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr"] Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.084786 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrc5x\" (UniqueName: \"kubernetes.io/projected/fd75fa45-6291-40b2-b40b-8763e6e72229-kube-api-access-lrc5x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.085400 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.085480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.187643 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.187741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.187822 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrc5x\" (UniqueName: \"kubernetes.io/projected/fd75fa45-6291-40b2-b40b-8763e6e72229-kube-api-access-lrc5x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.194290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.194292 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.208533 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrc5x\" (UniqueName: \"kubernetes.io/projected/fd75fa45-6291-40b2-b40b-8763e6e72229-kube-api-access-lrc5x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.328569 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.894681 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr"] Dec 10 19:27:08 crc kubenswrapper[4894]: I1210 19:27:08.903605 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" event={"ID":"fd75fa45-6291-40b2-b40b-8763e6e72229","Type":"ContainerStarted","Data":"d036aa79f9d72e5b5f3529a2f9748264fa70c71f7a45c276c6e7d98325052e22"} Dec 10 19:27:09 crc kubenswrapper[4894]: I1210 19:27:09.913250 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" event={"ID":"fd75fa45-6291-40b2-b40b-8763e6e72229","Type":"ContainerStarted","Data":"61d9188b5a0cee6f3b0352e186d08d359593413323eea9a8704f8432b7ae405b"} Dec 10 19:27:15 crc kubenswrapper[4894]: I1210 19:27:15.023613 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" podStartSLOduration=7.509116107 podStartE2EDuration="8.023590979s" podCreationTimestamp="2025-12-10 19:27:07 +0000 UTC" firstStartedPulling="2025-12-10 19:27:08.892440836 +0000 UTC m=+1911.687288638" lastFinishedPulling="2025-12-10 19:27:09.406915708 +0000 UTC m=+1912.201763510" observedRunningTime="2025-12-10 19:27:09.944995987 +0000 UTC m=+1912.739843779" watchObservedRunningTime="2025-12-10 19:27:15.023590979 +0000 UTC m=+1917.818438761" Dec 10 19:27:15 crc kubenswrapper[4894]: I1210 19:27:15.030796 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b2w4f"] Dec 10 19:27:15 crc kubenswrapper[4894]: I1210 19:27:15.040094 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b2w4f"] Dec 10 19:27:15 crc kubenswrapper[4894]: I1210 19:27:15.553140 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aac2573-2b49-4dcf-b808-d1adc6435bc9" path="/var/lib/kubelet/pods/8aac2573-2b49-4dcf-b808-d1adc6435bc9/volumes" Dec 10 19:27:17 crc kubenswrapper[4894]: I1210 19:27:17.559791 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:27:17 crc kubenswrapper[4894]: E1210 19:27:17.565135 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:27:18 crc kubenswrapper[4894]: I1210 19:27:18.035011 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5765m"] Dec 10 19:27:18 crc kubenswrapper[4894]: I1210 19:27:18.047171 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5765m"] Dec 10 19:27:19 crc kubenswrapper[4894]: I1210 19:27:19.035838 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-5bq9r"] Dec 10 19:27:19 crc kubenswrapper[4894]: I1210 19:27:19.046175 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-5bq9r"] Dec 10 19:27:19 crc kubenswrapper[4894]: I1210 19:27:19.558665 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6524d3b3-56ee-4a03-84a5-dfa9c4c27926" path="/var/lib/kubelet/pods/6524d3b3-56ee-4a03-84a5-dfa9c4c27926/volumes" Dec 10 19:27:19 crc kubenswrapper[4894]: I1210 19:27:19.559908 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bedfd09b-1fdb-43e9-88d7-47fd25103af6" path="/var/lib/kubelet/pods/bedfd09b-1fdb-43e9-88d7-47fd25103af6/volumes" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.531015 4894 scope.go:117] "RemoveContainer" containerID="c47cce434546c49a726e3eb930414b2f4c60861ec018316d9ce56d6b7190793a" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.584241 4894 scope.go:117] "RemoveContainer" containerID="6eeddc13ccc3804f8d502f50d581d43f84041df7a038c0e510a577216b5c622b" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.629198 4894 scope.go:117] "RemoveContainer" containerID="5c683b8f14854e3cedf5e129dd71f810dad9575381ca6f19aecd8e7f81cc44e7" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.684170 4894 scope.go:117] "RemoveContainer" containerID="a4109b1867a2f842962e0132f81de7089e0ec454657147bb1b03313921a51eb2" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.719718 4894 scope.go:117] "RemoveContainer" containerID="199379249cc0934c4aaa80bdb58285e0403aecb4e7f339830f8d02331bf82029" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.769285 4894 scope.go:117] "RemoveContainer" containerID="dc891aa04e9236ae856c527ec99f570d500cf5af2ff64596115282fa028072e9" Dec 10 19:27:21 crc kubenswrapper[4894]: I1210 19:27:21.812636 4894 scope.go:117] "RemoveContainer" containerID="1b334d8718fa9f84010e45a2ba7dbb672f710af8d56ea574e4b19198bd4f08a5" Dec 10 19:27:28 crc kubenswrapper[4894]: I1210 19:27:28.541418 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:27:29 crc kubenswrapper[4894]: I1210 19:27:29.107748 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"1a2a45ddb792c0c33a3e6ec306dbe31ee0324d8b7af5e3d8fe9603dedc9d4533"} Dec 10 19:28:11 crc kubenswrapper[4894]: I1210 19:28:11.052747 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6d03-account-create-update-gcqt5"] Dec 10 19:28:11 crc kubenswrapper[4894]: I1210 19:28:11.065712 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6d03-account-create-update-gcqt5"] Dec 10 19:28:11 crc kubenswrapper[4894]: I1210 19:28:11.561885 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93744ff6-5842-4ea1-9d60-0859da1c5dc1" path="/var/lib/kubelet/pods/93744ff6-5842-4ea1-9d60-0859da1c5dc1/volumes" Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.049796 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-54be-account-create-update-lxj4w"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.068220 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-w726g"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.079414 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-810e-account-create-update-b6bk8"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.087358 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-zc6w6"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.096955 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-54be-account-create-update-lxj4w"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.112352 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-w726g"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.123887 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-810e-account-create-update-b6bk8"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.129732 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-zc6w6"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.142180 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kmkpd"] Dec 10 19:28:12 crc kubenswrapper[4894]: I1210 19:28:12.153748 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kmkpd"] Dec 10 19:28:13 crc kubenswrapper[4894]: I1210 19:28:13.572166 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11d93ec8-1df6-456e-9043-8852cc63a87d" path="/var/lib/kubelet/pods/11d93ec8-1df6-456e-9043-8852cc63a87d/volumes" Dec 10 19:28:13 crc kubenswrapper[4894]: I1210 19:28:13.574331 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33f598ed-63ab-4a69-acb3-33fe9b9fabd8" path="/var/lib/kubelet/pods/33f598ed-63ab-4a69-acb3-33fe9b9fabd8/volumes" Dec 10 19:28:13 crc kubenswrapper[4894]: I1210 19:28:13.575352 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="607bf25a-d450-439f-977a-2f38c41f017f" path="/var/lib/kubelet/pods/607bf25a-d450-439f-977a-2f38c41f017f/volumes" Dec 10 19:28:13 crc kubenswrapper[4894]: I1210 19:28:13.576285 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad584c11-9601-4710-b5d4-fc6e58841d4f" path="/var/lib/kubelet/pods/ad584c11-9601-4710-b5d4-fc6e58841d4f/volumes" Dec 10 19:28:13 crc kubenswrapper[4894]: I1210 19:28:13.578172 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5ba7db9-ba2c-4cb7-9edc-2ed985264548" path="/var/lib/kubelet/pods/e5ba7db9-ba2c-4cb7-9edc-2ed985264548/volumes" Dec 10 19:28:22 crc kubenswrapper[4894]: I1210 19:28:22.000691 4894 scope.go:117] "RemoveContainer" containerID="d3ccec34a613b7f6d82204c1d1f81ef7f348e9e33fee501e976454437c92deff" Dec 10 19:28:22 crc kubenswrapper[4894]: I1210 19:28:22.026883 4894 scope.go:117] "RemoveContainer" containerID="2bef1b253b1b5483e4fd58ba310bf18366e3ff5174d13d377696dc421db3f0ec" Dec 10 19:28:22 crc kubenswrapper[4894]: I1210 19:28:22.101948 4894 scope.go:117] "RemoveContainer" containerID="e8b544c7dd401ce2ba5607d298b1d009e44cd9de07eed07c0ce0788cc29530fd" Dec 10 19:28:22 crc kubenswrapper[4894]: I1210 19:28:22.139380 4894 scope.go:117] "RemoveContainer" containerID="e2e992155b59dcd2f5f248a8d687f2d3d01529b6648fcb5327966142cbb345f3" Dec 10 19:28:22 crc kubenswrapper[4894]: I1210 19:28:22.184057 4894 scope.go:117] "RemoveContainer" containerID="b5da647cf904e8ce6e2a76e4dfe990f116ce27b895b253af3d8122534975e540" Dec 10 19:28:22 crc kubenswrapper[4894]: I1210 19:28:22.236317 4894 scope.go:117] "RemoveContainer" containerID="e6ea07194db43ce02d28708a77667b104f90088b218e1a94ff0e0526c7a8001f" Dec 10 19:28:26 crc kubenswrapper[4894]: I1210 19:28:26.754207 4894 generic.go:334] "Generic (PLEG): container finished" podID="fd75fa45-6291-40b2-b40b-8763e6e72229" containerID="61d9188b5a0cee6f3b0352e186d08d359593413323eea9a8704f8432b7ae405b" exitCode=0 Dec 10 19:28:26 crc kubenswrapper[4894]: I1210 19:28:26.754300 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" event={"ID":"fd75fa45-6291-40b2-b40b-8763e6e72229","Type":"ContainerDied","Data":"61d9188b5a0cee6f3b0352e186d08d359593413323eea9a8704f8432b7ae405b"} Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.372655 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.429691 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-ssh-key\") pod \"fd75fa45-6291-40b2-b40b-8763e6e72229\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.429798 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrc5x\" (UniqueName: \"kubernetes.io/projected/fd75fa45-6291-40b2-b40b-8763e6e72229-kube-api-access-lrc5x\") pod \"fd75fa45-6291-40b2-b40b-8763e6e72229\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.430008 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-inventory\") pod \"fd75fa45-6291-40b2-b40b-8763e6e72229\" (UID: \"fd75fa45-6291-40b2-b40b-8763e6e72229\") " Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.435050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd75fa45-6291-40b2-b40b-8763e6e72229-kube-api-access-lrc5x" (OuterVolumeSpecName: "kube-api-access-lrc5x") pod "fd75fa45-6291-40b2-b40b-8763e6e72229" (UID: "fd75fa45-6291-40b2-b40b-8763e6e72229"). InnerVolumeSpecName "kube-api-access-lrc5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.460883 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd75fa45-6291-40b2-b40b-8763e6e72229" (UID: "fd75fa45-6291-40b2-b40b-8763e6e72229"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.466084 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-inventory" (OuterVolumeSpecName: "inventory") pod "fd75fa45-6291-40b2-b40b-8763e6e72229" (UID: "fd75fa45-6291-40b2-b40b-8763e6e72229"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.532904 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.532938 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrc5x\" (UniqueName: \"kubernetes.io/projected/fd75fa45-6291-40b2-b40b-8763e6e72229-kube-api-access-lrc5x\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.532974 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd75fa45-6291-40b2-b40b-8763e6e72229-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.781568 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" event={"ID":"fd75fa45-6291-40b2-b40b-8763e6e72229","Type":"ContainerDied","Data":"d036aa79f9d72e5b5f3529a2f9748264fa70c71f7a45c276c6e7d98325052e22"} Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.782109 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d036aa79f9d72e5b5f3529a2f9748264fa70c71f7a45c276c6e7d98325052e22" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.781622 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.893817 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m"] Dec 10 19:28:28 crc kubenswrapper[4894]: E1210 19:28:28.894301 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd75fa45-6291-40b2-b40b-8763e6e72229" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.894322 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd75fa45-6291-40b2-b40b-8763e6e72229" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.894546 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd75fa45-6291-40b2-b40b-8763e6e72229" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.895325 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.906192 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m"] Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.940430 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.940493 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw572\" (UniqueName: \"kubernetes.io/projected/7a4c1129-0acd-4a3e-a258-281efc367831-kube-api-access-cw572\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.940630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.943375 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.943425 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.943598 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:28:28 crc kubenswrapper[4894]: I1210 19:28:28.944492 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.042805 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.043040 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.043101 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw572\" (UniqueName: \"kubernetes.io/projected/7a4c1129-0acd-4a3e-a258-281efc367831-kube-api-access-cw572\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.047603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.050500 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.062597 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw572\" (UniqueName: \"kubernetes.io/projected/7a4c1129-0acd-4a3e-a258-281efc367831-kube-api-access-cw572\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ht74m\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.265566 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:29 crc kubenswrapper[4894]: I1210 19:28:29.834027 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m"] Dec 10 19:28:30 crc kubenswrapper[4894]: I1210 19:28:30.810033 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" event={"ID":"7a4c1129-0acd-4a3e-a258-281efc367831","Type":"ContainerStarted","Data":"5740fd09741650f3d86d6615fe0560b0bf1576d5b7d030a7959d6b58b0ffcf3a"} Dec 10 19:28:31 crc kubenswrapper[4894]: I1210 19:28:31.821628 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" event={"ID":"7a4c1129-0acd-4a3e-a258-281efc367831","Type":"ContainerStarted","Data":"c67e9d26ee00177d0f8d36c00e128fb5f1b5fd4663d0331df14a2d2233848ccd"} Dec 10 19:28:31 crc kubenswrapper[4894]: I1210 19:28:31.858214 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" podStartSLOduration=2.9875830629999998 podStartE2EDuration="3.858193947s" podCreationTimestamp="2025-12-10 19:28:28 +0000 UTC" firstStartedPulling="2025-12-10 19:28:29.841887036 +0000 UTC m=+1992.636734828" lastFinishedPulling="2025-12-10 19:28:30.71249792 +0000 UTC m=+1993.507345712" observedRunningTime="2025-12-10 19:28:31.841155451 +0000 UTC m=+1994.636003273" watchObservedRunningTime="2025-12-10 19:28:31.858193947 +0000 UTC m=+1994.653041739" Dec 10 19:28:36 crc kubenswrapper[4894]: I1210 19:28:36.878624 4894 generic.go:334] "Generic (PLEG): container finished" podID="7a4c1129-0acd-4a3e-a258-281efc367831" containerID="c67e9d26ee00177d0f8d36c00e128fb5f1b5fd4663d0331df14a2d2233848ccd" exitCode=0 Dec 10 19:28:36 crc kubenswrapper[4894]: I1210 19:28:36.878710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" event={"ID":"7a4c1129-0acd-4a3e-a258-281efc367831","Type":"ContainerDied","Data":"c67e9d26ee00177d0f8d36c00e128fb5f1b5fd4663d0331df14a2d2233848ccd"} Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.375512 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.457395 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-ssh-key\") pod \"7a4c1129-0acd-4a3e-a258-281efc367831\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.457910 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-inventory\") pod \"7a4c1129-0acd-4a3e-a258-281efc367831\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.458145 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw572\" (UniqueName: \"kubernetes.io/projected/7a4c1129-0acd-4a3e-a258-281efc367831-kube-api-access-cw572\") pod \"7a4c1129-0acd-4a3e-a258-281efc367831\" (UID: \"7a4c1129-0acd-4a3e-a258-281efc367831\") " Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.470991 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a4c1129-0acd-4a3e-a258-281efc367831-kube-api-access-cw572" (OuterVolumeSpecName: "kube-api-access-cw572") pod "7a4c1129-0acd-4a3e-a258-281efc367831" (UID: "7a4c1129-0acd-4a3e-a258-281efc367831"). InnerVolumeSpecName "kube-api-access-cw572". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.485445 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7a4c1129-0acd-4a3e-a258-281efc367831" (UID: "7a4c1129-0acd-4a3e-a258-281efc367831"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.486416 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-inventory" (OuterVolumeSpecName: "inventory") pod "7a4c1129-0acd-4a3e-a258-281efc367831" (UID: "7a4c1129-0acd-4a3e-a258-281efc367831"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.560248 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw572\" (UniqueName: \"kubernetes.io/projected/7a4c1129-0acd-4a3e-a258-281efc367831-kube-api-access-cw572\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.560441 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.560530 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a4c1129-0acd-4a3e-a258-281efc367831-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.897877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" event={"ID":"7a4c1129-0acd-4a3e-a258-281efc367831","Type":"ContainerDied","Data":"5740fd09741650f3d86d6615fe0560b0bf1576d5b7d030a7959d6b58b0ffcf3a"} Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.897925 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5740fd09741650f3d86d6615fe0560b0bf1576d5b7d030a7959d6b58b0ffcf3a" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.898113 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ht74m" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.987760 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m"] Dec 10 19:28:38 crc kubenswrapper[4894]: E1210 19:28:38.988203 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a4c1129-0acd-4a3e-a258-281efc367831" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.988218 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a4c1129-0acd-4a3e-a258-281efc367831" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.988454 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a4c1129-0acd-4a3e-a258-281efc367831" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.989219 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.997175 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m"] Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.997668 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.997892 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.998305 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:28:38 crc kubenswrapper[4894]: I1210 19:28:38.998407 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.071952 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6md9r\" (UniqueName: \"kubernetes.io/projected/2583f762-94a3-4dab-9ace-2e1933c0af69-kube-api-access-6md9r\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.072044 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.072095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.174163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.174249 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.174560 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6md9r\" (UniqueName: \"kubernetes.io/projected/2583f762-94a3-4dab-9ace-2e1933c0af69-kube-api-access-6md9r\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.179447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.190370 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.201758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6md9r\" (UniqueName: \"kubernetes.io/projected/2583f762-94a3-4dab-9ace-2e1933c0af69-kube-api-access-6md9r\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vh26m\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.324276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.883676 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m"] Dec 10 19:28:39 crc kubenswrapper[4894]: I1210 19:28:39.909077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" event={"ID":"2583f762-94a3-4dab-9ace-2e1933c0af69","Type":"ContainerStarted","Data":"5be9d2fe76028983287935c82b78cbea247601bcee79ed10ab2d6ff7540940d6"} Dec 10 19:28:40 crc kubenswrapper[4894]: I1210 19:28:40.921212 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" event={"ID":"2583f762-94a3-4dab-9ace-2e1933c0af69","Type":"ContainerStarted","Data":"4a3b47585b4ccf9ba3a9d3986a961738de8c7e01e44ec7d4c0ddf9245adcc2df"} Dec 10 19:28:40 crc kubenswrapper[4894]: I1210 19:28:40.948661 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" podStartSLOduration=2.521937213 podStartE2EDuration="2.948638765s" podCreationTimestamp="2025-12-10 19:28:38 +0000 UTC" firstStartedPulling="2025-12-10 19:28:39.89590387 +0000 UTC m=+2002.690751662" lastFinishedPulling="2025-12-10 19:28:40.322605422 +0000 UTC m=+2003.117453214" observedRunningTime="2025-12-10 19:28:40.937954293 +0000 UTC m=+2003.732802105" watchObservedRunningTime="2025-12-10 19:28:40.948638765 +0000 UTC m=+2003.743486567" Dec 10 19:28:42 crc kubenswrapper[4894]: I1210 19:28:42.041923 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k84gv"] Dec 10 19:28:42 crc kubenswrapper[4894]: I1210 19:28:42.056648 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k84gv"] Dec 10 19:28:43 crc kubenswrapper[4894]: I1210 19:28:43.554512 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a93527-c2bb-4968-862c-02c60d6fe93c" path="/var/lib/kubelet/pods/53a93527-c2bb-4968-862c-02c60d6fe93c/volumes" Dec 10 19:29:03 crc kubenswrapper[4894]: I1210 19:29:03.055407 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-pglr9"] Dec 10 19:29:03 crc kubenswrapper[4894]: I1210 19:29:03.068270 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-pglr9"] Dec 10 19:29:03 crc kubenswrapper[4894]: I1210 19:29:03.560059 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57e8fe0-db28-42be-aa0d-f39f751d7897" path="/var/lib/kubelet/pods/a57e8fe0-db28-42be-aa0d-f39f751d7897/volumes" Dec 10 19:29:07 crc kubenswrapper[4894]: I1210 19:29:07.028055 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8ntxv"] Dec 10 19:29:07 crc kubenswrapper[4894]: I1210 19:29:07.038053 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8ntxv"] Dec 10 19:29:07 crc kubenswrapper[4894]: I1210 19:29:07.555204 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="896eec35-0551-4bf5-9afa-179674319fdd" path="/var/lib/kubelet/pods/896eec35-0551-4bf5-9afa-179674319fdd/volumes" Dec 10 19:29:22 crc kubenswrapper[4894]: I1210 19:29:22.401731 4894 scope.go:117] "RemoveContainer" containerID="f60fb0e5527c8ea654ab78a3f1641d06d5e2890ada4ec13c23585124cea12666" Dec 10 19:29:22 crc kubenswrapper[4894]: I1210 19:29:22.484311 4894 scope.go:117] "RemoveContainer" containerID="02f90fc7a4bd62f67dfd1457046169268063d085c23f8c34c4abcbe11a807ed8" Dec 10 19:29:22 crc kubenswrapper[4894]: I1210 19:29:22.536660 4894 scope.go:117] "RemoveContainer" containerID="c438f8d74dcb947bced6be6cd664e5c1aa2a1043d5498dc2dcf32d572c55c989" Dec 10 19:29:23 crc kubenswrapper[4894]: I1210 19:29:23.753415 4894 generic.go:334] "Generic (PLEG): container finished" podID="2583f762-94a3-4dab-9ace-2e1933c0af69" containerID="4a3b47585b4ccf9ba3a9d3986a961738de8c7e01e44ec7d4c0ddf9245adcc2df" exitCode=0 Dec 10 19:29:23 crc kubenswrapper[4894]: I1210 19:29:23.753499 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" event={"ID":"2583f762-94a3-4dab-9ace-2e1933c0af69","Type":"ContainerDied","Data":"4a3b47585b4ccf9ba3a9d3986a961738de8c7e01e44ec7d4c0ddf9245adcc2df"} Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.263453 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.375571 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6md9r\" (UniqueName: \"kubernetes.io/projected/2583f762-94a3-4dab-9ace-2e1933c0af69-kube-api-access-6md9r\") pod \"2583f762-94a3-4dab-9ace-2e1933c0af69\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.375955 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-ssh-key\") pod \"2583f762-94a3-4dab-9ace-2e1933c0af69\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.376295 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-inventory\") pod \"2583f762-94a3-4dab-9ace-2e1933c0af69\" (UID: \"2583f762-94a3-4dab-9ace-2e1933c0af69\") " Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.382989 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2583f762-94a3-4dab-9ace-2e1933c0af69-kube-api-access-6md9r" (OuterVolumeSpecName: "kube-api-access-6md9r") pod "2583f762-94a3-4dab-9ace-2e1933c0af69" (UID: "2583f762-94a3-4dab-9ace-2e1933c0af69"). InnerVolumeSpecName "kube-api-access-6md9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.409280 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-inventory" (OuterVolumeSpecName: "inventory") pod "2583f762-94a3-4dab-9ace-2e1933c0af69" (UID: "2583f762-94a3-4dab-9ace-2e1933c0af69"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.425552 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2583f762-94a3-4dab-9ace-2e1933c0af69" (UID: "2583f762-94a3-4dab-9ace-2e1933c0af69"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.479377 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.479438 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6md9r\" (UniqueName: \"kubernetes.io/projected/2583f762-94a3-4dab-9ace-2e1933c0af69-kube-api-access-6md9r\") on node \"crc\" DevicePath \"\"" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.479450 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2583f762-94a3-4dab-9ace-2e1933c0af69-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.776598 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" event={"ID":"2583f762-94a3-4dab-9ace-2e1933c0af69","Type":"ContainerDied","Data":"5be9d2fe76028983287935c82b78cbea247601bcee79ed10ab2d6ff7540940d6"} Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.776647 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5be9d2fe76028983287935c82b78cbea247601bcee79ed10ab2d6ff7540940d6" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.776683 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vh26m" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.869187 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft"] Dec 10 19:29:25 crc kubenswrapper[4894]: E1210 19:29:25.870526 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2583f762-94a3-4dab-9ace-2e1933c0af69" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.870547 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2583f762-94a3-4dab-9ace-2e1933c0af69" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.870742 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2583f762-94a3-4dab-9ace-2e1933c0af69" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.871471 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.876755 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.876968 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.877153 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.877264 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.885305 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft"] Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.992237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.992279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:25 crc kubenswrapper[4894]: I1210 19:29:25.992371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxp85\" (UniqueName: \"kubernetes.io/projected/bc921db2-6d84-4c34-9de5-12966910f500-kube-api-access-dxp85\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.094040 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.094108 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.094224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxp85\" (UniqueName: \"kubernetes.io/projected/bc921db2-6d84-4c34-9de5-12966910f500-kube-api-access-dxp85\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.099058 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.099997 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.123837 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxp85\" (UniqueName: \"kubernetes.io/projected/bc921db2-6d84-4c34-9de5-12966910f500-kube-api-access-dxp85\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-597ft\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.213032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:29:26 crc kubenswrapper[4894]: I1210 19:29:26.822926 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft"] Dec 10 19:29:27 crc kubenswrapper[4894]: I1210 19:29:27.808481 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" event={"ID":"bc921db2-6d84-4c34-9de5-12966910f500","Type":"ContainerStarted","Data":"2cf222abd7bf400cae8f70ebded8db2afb966af4e5336da3b23d321a33bc7c02"} Dec 10 19:29:27 crc kubenswrapper[4894]: I1210 19:29:27.809301 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" event={"ID":"bc921db2-6d84-4c34-9de5-12966910f500","Type":"ContainerStarted","Data":"1ff27bc5ac4f56d2951af68e63863b4624366a94cc3338226b9db4461cd08720"} Dec 10 19:29:27 crc kubenswrapper[4894]: I1210 19:29:27.839577 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" podStartSLOduration=2.172925282 podStartE2EDuration="2.839555066s" podCreationTimestamp="2025-12-10 19:29:25 +0000 UTC" firstStartedPulling="2025-12-10 19:29:26.833159338 +0000 UTC m=+2049.628007140" lastFinishedPulling="2025-12-10 19:29:27.499789132 +0000 UTC m=+2050.294636924" observedRunningTime="2025-12-10 19:29:27.826945211 +0000 UTC m=+2050.621793073" watchObservedRunningTime="2025-12-10 19:29:27.839555066 +0000 UTC m=+2050.634402868" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.141581 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n47z7"] Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.156300 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.179119 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n47z7"] Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.203513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s4mv\" (UniqueName: \"kubernetes.io/projected/06f4c78a-52c7-48c1-be8c-0fbfa4859392-kube-api-access-6s4mv\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.204333 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-utilities\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.205095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-catalog-content\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.306785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s4mv\" (UniqueName: \"kubernetes.io/projected/06f4c78a-52c7-48c1-be8c-0fbfa4859392-kube-api-access-6s4mv\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.306834 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-utilities\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.306994 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-catalog-content\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.307449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-utilities\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.307553 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-catalog-content\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.342928 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s4mv\" (UniqueName: \"kubernetes.io/projected/06f4c78a-52c7-48c1-be8c-0fbfa4859392-kube-api-access-6s4mv\") pod \"redhat-operators-n47z7\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:44 crc kubenswrapper[4894]: I1210 19:29:44.518814 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:45 crc kubenswrapper[4894]: I1210 19:29:45.054587 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n47z7"] Dec 10 19:29:45 crc kubenswrapper[4894]: W1210 19:29:45.060105 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06f4c78a_52c7_48c1_be8c_0fbfa4859392.slice/crio-b2a96304658ee6aeca1b03c4db44fbc661f1affe4533e502bb65d622eb943a4a WatchSource:0}: Error finding container b2a96304658ee6aeca1b03c4db44fbc661f1affe4533e502bb65d622eb943a4a: Status 404 returned error can't find the container with id b2a96304658ee6aeca1b03c4db44fbc661f1affe4533e502bb65d622eb943a4a Dec 10 19:29:45 crc kubenswrapper[4894]: I1210 19:29:45.987199 4894 generic.go:334] "Generic (PLEG): container finished" podID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerID="ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501" exitCode=0 Dec 10 19:29:45 crc kubenswrapper[4894]: I1210 19:29:45.987523 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerDied","Data":"ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501"} Dec 10 19:29:45 crc kubenswrapper[4894]: I1210 19:29:45.987569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerStarted","Data":"b2a96304658ee6aeca1b03c4db44fbc661f1affe4533e502bb65d622eb943a4a"} Dec 10 19:29:47 crc kubenswrapper[4894]: I1210 19:29:47.051162 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9pwpt"] Dec 10 19:29:47 crc kubenswrapper[4894]: I1210 19:29:47.061385 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9pwpt"] Dec 10 19:29:47 crc kubenswrapper[4894]: I1210 19:29:47.558601 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4" path="/var/lib/kubelet/pods/2478d7d0-05d0-4dbc-b4e3-d500b43f2ec4/volumes" Dec 10 19:29:48 crc kubenswrapper[4894]: I1210 19:29:48.032838 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerStarted","Data":"a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89"} Dec 10 19:29:49 crc kubenswrapper[4894]: I1210 19:29:49.577644 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:29:49 crc kubenswrapper[4894]: I1210 19:29:49.578130 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:29:51 crc kubenswrapper[4894]: I1210 19:29:51.064605 4894 generic.go:334] "Generic (PLEG): container finished" podID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerID="a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89" exitCode=0 Dec 10 19:29:51 crc kubenswrapper[4894]: I1210 19:29:51.064708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerDied","Data":"a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89"} Dec 10 19:29:53 crc kubenswrapper[4894]: I1210 19:29:53.111510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerStarted","Data":"245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425"} Dec 10 19:29:53 crc kubenswrapper[4894]: I1210 19:29:53.132352 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n47z7" podStartSLOduration=3.019440985 podStartE2EDuration="9.132330969s" podCreationTimestamp="2025-12-10 19:29:44 +0000 UTC" firstStartedPulling="2025-12-10 19:29:45.993111533 +0000 UTC m=+2068.787959325" lastFinishedPulling="2025-12-10 19:29:52.106001517 +0000 UTC m=+2074.900849309" observedRunningTime="2025-12-10 19:29:53.128080953 +0000 UTC m=+2075.922928775" watchObservedRunningTime="2025-12-10 19:29:53.132330969 +0000 UTC m=+2075.927178771" Dec 10 19:29:54 crc kubenswrapper[4894]: I1210 19:29:54.519447 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:54 crc kubenswrapper[4894]: I1210 19:29:54.519785 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:29:55 crc kubenswrapper[4894]: I1210 19:29:55.578178 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n47z7" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="registry-server" probeResult="failure" output=< Dec 10 19:29:55 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:29:55 crc kubenswrapper[4894]: > Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.143517 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2"] Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.145309 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.149767 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.150660 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.152680 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2"] Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.240407 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s7jd\" (UniqueName: \"kubernetes.io/projected/f6bdce4b-fa97-46a4-9293-b314220e0ebe-kube-api-access-9s7jd\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.240466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6bdce4b-fa97-46a4-9293-b314220e0ebe-config-volume\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.240694 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6bdce4b-fa97-46a4-9293-b314220e0ebe-secret-volume\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.342326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6bdce4b-fa97-46a4-9293-b314220e0ebe-secret-volume\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.342463 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s7jd\" (UniqueName: \"kubernetes.io/projected/f6bdce4b-fa97-46a4-9293-b314220e0ebe-kube-api-access-9s7jd\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.342487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6bdce4b-fa97-46a4-9293-b314220e0ebe-config-volume\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.343389 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6bdce4b-fa97-46a4-9293-b314220e0ebe-config-volume\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.348099 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6bdce4b-fa97-46a4-9293-b314220e0ebe-secret-volume\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.359325 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s7jd\" (UniqueName: \"kubernetes.io/projected/f6bdce4b-fa97-46a4-9293-b314220e0ebe-kube-api-access-9s7jd\") pod \"collect-profiles-29423250-987w2\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.476073 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:00 crc kubenswrapper[4894]: I1210 19:30:00.945757 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2"] Dec 10 19:30:01 crc kubenswrapper[4894]: I1210 19:30:01.187131 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" event={"ID":"f6bdce4b-fa97-46a4-9293-b314220e0ebe","Type":"ContainerStarted","Data":"9a69acbdb4a6b34cc0b7484bcc6c7068403b314e5f6b4175488013bfdd3b2517"} Dec 10 19:30:01 crc kubenswrapper[4894]: I1210 19:30:01.187184 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" event={"ID":"f6bdce4b-fa97-46a4-9293-b314220e0ebe","Type":"ContainerStarted","Data":"f47acd23a82d0848d44dc541e3448c1aa5c1dc4ebfa1d633cec908328996798e"} Dec 10 19:30:01 crc kubenswrapper[4894]: I1210 19:30:01.210332 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" podStartSLOduration=1.210310613 podStartE2EDuration="1.210310613s" podCreationTimestamp="2025-12-10 19:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:30:01.209334767 +0000 UTC m=+2084.004182569" watchObservedRunningTime="2025-12-10 19:30:01.210310613 +0000 UTC m=+2084.005158415" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.197161 4894 generic.go:334] "Generic (PLEG): container finished" podID="f6bdce4b-fa97-46a4-9293-b314220e0ebe" containerID="9a69acbdb4a6b34cc0b7484bcc6c7068403b314e5f6b4175488013bfdd3b2517" exitCode=0 Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.197225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" event={"ID":"f6bdce4b-fa97-46a4-9293-b314220e0ebe","Type":"ContainerDied","Data":"9a69acbdb4a6b34cc0b7484bcc6c7068403b314e5f6b4175488013bfdd3b2517"} Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.620811 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nx7f7"] Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.623486 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.644778 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nx7f7"] Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.689019 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-utilities\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.689068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-catalog-content\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.689140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv5kt\" (UniqueName: \"kubernetes.io/projected/de46bbe8-3eca-402c-b0ec-d41f4a70a469-kube-api-access-qv5kt\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.790678 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv5kt\" (UniqueName: \"kubernetes.io/projected/de46bbe8-3eca-402c-b0ec-d41f4a70a469-kube-api-access-qv5kt\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.790868 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-utilities\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.790900 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-catalog-content\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.791415 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-catalog-content\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.791522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-utilities\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.810612 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv5kt\" (UniqueName: \"kubernetes.io/projected/de46bbe8-3eca-402c-b0ec-d41f4a70a469-kube-api-access-qv5kt\") pod \"redhat-marketplace-nx7f7\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:02 crc kubenswrapper[4894]: I1210 19:30:02.978235 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.463811 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nx7f7"] Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.593683 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.710580 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6bdce4b-fa97-46a4-9293-b314220e0ebe-secret-volume\") pod \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.710764 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6bdce4b-fa97-46a4-9293-b314220e0ebe-config-volume\") pod \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.710888 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s7jd\" (UniqueName: \"kubernetes.io/projected/f6bdce4b-fa97-46a4-9293-b314220e0ebe-kube-api-access-9s7jd\") pod \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\" (UID: \"f6bdce4b-fa97-46a4-9293-b314220e0ebe\") " Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.711429 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6bdce4b-fa97-46a4-9293-b314220e0ebe-config-volume" (OuterVolumeSpecName: "config-volume") pod "f6bdce4b-fa97-46a4-9293-b314220e0ebe" (UID: "f6bdce4b-fa97-46a4-9293-b314220e0ebe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.712006 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6bdce4b-fa97-46a4-9293-b314220e0ebe-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.717446 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6bdce4b-fa97-46a4-9293-b314220e0ebe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f6bdce4b-fa97-46a4-9293-b314220e0ebe" (UID: "f6bdce4b-fa97-46a4-9293-b314220e0ebe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.720117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6bdce4b-fa97-46a4-9293-b314220e0ebe-kube-api-access-9s7jd" (OuterVolumeSpecName: "kube-api-access-9s7jd") pod "f6bdce4b-fa97-46a4-9293-b314220e0ebe" (UID: "f6bdce4b-fa97-46a4-9293-b314220e0ebe"). InnerVolumeSpecName "kube-api-access-9s7jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.813296 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s7jd\" (UniqueName: \"kubernetes.io/projected/f6bdce4b-fa97-46a4-9293-b314220e0ebe-kube-api-access-9s7jd\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:03 crc kubenswrapper[4894]: I1210 19:30:03.813334 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6bdce4b-fa97-46a4-9293-b314220e0ebe-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.216521 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" event={"ID":"f6bdce4b-fa97-46a4-9293-b314220e0ebe","Type":"ContainerDied","Data":"f47acd23a82d0848d44dc541e3448c1aa5c1dc4ebfa1d633cec908328996798e"} Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.216571 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f47acd23a82d0848d44dc541e3448c1aa5c1dc4ebfa1d633cec908328996798e" Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.216634 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-987w2" Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.219926 4894 generic.go:334] "Generic (PLEG): container finished" podID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerID="a5ab5de61c62c1417a8aff6ce784c0af24c2a8908a0b2a526948c3a6fbc7c283" exitCode=0 Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.219968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerDied","Data":"a5ab5de61c62c1417a8aff6ce784c0af24c2a8908a0b2a526948c3a6fbc7c283"} Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.219994 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerStarted","Data":"870638010f19691a1d38a5387d9baff0490f2c6da41d5d50312d19ed16ae6268"} Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.323342 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n"] Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.332299 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-vpp6n"] Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.579019 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:30:04 crc kubenswrapper[4894]: I1210 19:30:04.632108 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:30:05 crc kubenswrapper[4894]: I1210 19:30:05.757354 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e39c4e75-61aa-4690-bff2-eb4be173b66e" path="/var/lib/kubelet/pods/e39c4e75-61aa-4690-bff2-eb4be173b66e/volumes" Dec 10 19:30:07 crc kubenswrapper[4894]: I1210 19:30:07.002335 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n47z7"] Dec 10 19:30:07 crc kubenswrapper[4894]: I1210 19:30:07.002879 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n47z7" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="registry-server" containerID="cri-o://245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425" gracePeriod=2 Dec 10 19:30:07 crc kubenswrapper[4894]: I1210 19:30:07.250706 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerStarted","Data":"29a7cf7f06bba0ab25f594e88e1ef9d5447382e4a77c6a0f4b2018f2477314c6"} Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.036524 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.086005 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s4mv\" (UniqueName: \"kubernetes.io/projected/06f4c78a-52c7-48c1-be8c-0fbfa4859392-kube-api-access-6s4mv\") pod \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.086240 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-utilities\") pod \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.086317 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-catalog-content\") pod \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\" (UID: \"06f4c78a-52c7-48c1-be8c-0fbfa4859392\") " Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.086967 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-utilities" (OuterVolumeSpecName: "utilities") pod "06f4c78a-52c7-48c1-be8c-0fbfa4859392" (UID: "06f4c78a-52c7-48c1-be8c-0fbfa4859392"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.101890 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f4c78a-52c7-48c1-be8c-0fbfa4859392-kube-api-access-6s4mv" (OuterVolumeSpecName: "kube-api-access-6s4mv") pod "06f4c78a-52c7-48c1-be8c-0fbfa4859392" (UID: "06f4c78a-52c7-48c1-be8c-0fbfa4859392"). InnerVolumeSpecName "kube-api-access-6s4mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.188707 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s4mv\" (UniqueName: \"kubernetes.io/projected/06f4c78a-52c7-48c1-be8c-0fbfa4859392-kube-api-access-6s4mv\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.188742 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.196215 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06f4c78a-52c7-48c1-be8c-0fbfa4859392" (UID: "06f4c78a-52c7-48c1-be8c-0fbfa4859392"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.262481 4894 generic.go:334] "Generic (PLEG): container finished" podID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerID="245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425" exitCode=0 Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.262657 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n47z7" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.262825 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerDied","Data":"245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425"} Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.263020 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n47z7" event={"ID":"06f4c78a-52c7-48c1-be8c-0fbfa4859392","Type":"ContainerDied","Data":"b2a96304658ee6aeca1b03c4db44fbc661f1affe4533e502bb65d622eb943a4a"} Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.263070 4894 scope.go:117] "RemoveContainer" containerID="245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.265225 4894 generic.go:334] "Generic (PLEG): container finished" podID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerID="29a7cf7f06bba0ab25f594e88e1ef9d5447382e4a77c6a0f4b2018f2477314c6" exitCode=0 Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.265263 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerDied","Data":"29a7cf7f06bba0ab25f594e88e1ef9d5447382e4a77c6a0f4b2018f2477314c6"} Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.290827 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f4c78a-52c7-48c1-be8c-0fbfa4859392-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.294836 4894 scope.go:117] "RemoveContainer" containerID="a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.331448 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n47z7"] Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.341489 4894 scope.go:117] "RemoveContainer" containerID="ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.344556 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n47z7"] Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.374340 4894 scope.go:117] "RemoveContainer" containerID="245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425" Dec 10 19:30:08 crc kubenswrapper[4894]: E1210 19:30:08.374656 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425\": container with ID starting with 245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425 not found: ID does not exist" containerID="245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.374687 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425"} err="failed to get container status \"245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425\": rpc error: code = NotFound desc = could not find container \"245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425\": container with ID starting with 245f2255ae9224c2200b6001c262ccce94d0b41090023193733e837144ec7425 not found: ID does not exist" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.374708 4894 scope.go:117] "RemoveContainer" containerID="a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89" Dec 10 19:30:08 crc kubenswrapper[4894]: E1210 19:30:08.375137 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89\": container with ID starting with a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89 not found: ID does not exist" containerID="a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.375170 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89"} err="failed to get container status \"a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89\": rpc error: code = NotFound desc = could not find container \"a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89\": container with ID starting with a8ae6a38b624f451af85970932b5f7cf6c0a906c488a5c8661cb3a5131332d89 not found: ID does not exist" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.375184 4894 scope.go:117] "RemoveContainer" containerID="ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501" Dec 10 19:30:08 crc kubenswrapper[4894]: E1210 19:30:08.375367 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501\": container with ID starting with ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501 not found: ID does not exist" containerID="ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501" Dec 10 19:30:08 crc kubenswrapper[4894]: I1210 19:30:08.375385 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501"} err="failed to get container status \"ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501\": rpc error: code = NotFound desc = could not find container \"ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501\": container with ID starting with ce5f5178e99dda5638f638304729390be2c90a79bdfc7f88bbfaf131dc7a2501 not found: ID does not exist" Dec 10 19:30:09 crc kubenswrapper[4894]: I1210 19:30:09.275400 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerStarted","Data":"6243213b26647481ee9adf47a47091cd3fa49239733c0e77da18526dee83ec60"} Dec 10 19:30:09 crc kubenswrapper[4894]: I1210 19:30:09.297643 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nx7f7" podStartSLOduration=2.796287581 podStartE2EDuration="7.297628504s" podCreationTimestamp="2025-12-10 19:30:02 +0000 UTC" firstStartedPulling="2025-12-10 19:30:04.221864107 +0000 UTC m=+2087.016711899" lastFinishedPulling="2025-12-10 19:30:08.72320503 +0000 UTC m=+2091.518052822" observedRunningTime="2025-12-10 19:30:09.293440579 +0000 UTC m=+2092.088288381" watchObservedRunningTime="2025-12-10 19:30:09.297628504 +0000 UTC m=+2092.092476286" Dec 10 19:30:09 crc kubenswrapper[4894]: I1210 19:30:09.564207 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" path="/var/lib/kubelet/pods/06f4c78a-52c7-48c1-be8c-0fbfa4859392/volumes" Dec 10 19:30:12 crc kubenswrapper[4894]: I1210 19:30:12.979275 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:12 crc kubenswrapper[4894]: I1210 19:30:12.979676 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:13 crc kubenswrapper[4894]: I1210 19:30:13.033614 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:13 crc kubenswrapper[4894]: I1210 19:30:13.377177 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:14 crc kubenswrapper[4894]: I1210 19:30:14.199724 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nx7f7"] Dec 10 19:30:15 crc kubenswrapper[4894]: I1210 19:30:15.333114 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nx7f7" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="registry-server" containerID="cri-o://6243213b26647481ee9adf47a47091cd3fa49239733c0e77da18526dee83ec60" gracePeriod=2 Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.343377 4894 generic.go:334] "Generic (PLEG): container finished" podID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerID="6243213b26647481ee9adf47a47091cd3fa49239733c0e77da18526dee83ec60" exitCode=0 Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.343510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerDied","Data":"6243213b26647481ee9adf47a47091cd3fa49239733c0e77da18526dee83ec60"} Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.451461 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.555656 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-catalog-content\") pod \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.555733 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-utilities\") pod \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.555824 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv5kt\" (UniqueName: \"kubernetes.io/projected/de46bbe8-3eca-402c-b0ec-d41f4a70a469-kube-api-access-qv5kt\") pod \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\" (UID: \"de46bbe8-3eca-402c-b0ec-d41f4a70a469\") " Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.556503 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-utilities" (OuterVolumeSpecName: "utilities") pod "de46bbe8-3eca-402c-b0ec-d41f4a70a469" (UID: "de46bbe8-3eca-402c-b0ec-d41f4a70a469"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.562080 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de46bbe8-3eca-402c-b0ec-d41f4a70a469-kube-api-access-qv5kt" (OuterVolumeSpecName: "kube-api-access-qv5kt") pod "de46bbe8-3eca-402c-b0ec-d41f4a70a469" (UID: "de46bbe8-3eca-402c-b0ec-d41f4a70a469"). InnerVolumeSpecName "kube-api-access-qv5kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.576729 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de46bbe8-3eca-402c-b0ec-d41f4a70a469" (UID: "de46bbe8-3eca-402c-b0ec-d41f4a70a469"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.659480 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.659714 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de46bbe8-3eca-402c-b0ec-d41f4a70a469-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:16 crc kubenswrapper[4894]: I1210 19:30:16.659806 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv5kt\" (UniqueName: \"kubernetes.io/projected/de46bbe8-3eca-402c-b0ec-d41f4a70a469-kube-api-access-qv5kt\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.354156 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nx7f7" event={"ID":"de46bbe8-3eca-402c-b0ec-d41f4a70a469","Type":"ContainerDied","Data":"870638010f19691a1d38a5387d9baff0490f2c6da41d5d50312d19ed16ae6268"} Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.354398 4894 scope.go:117] "RemoveContainer" containerID="6243213b26647481ee9adf47a47091cd3fa49239733c0e77da18526dee83ec60" Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.354215 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nx7f7" Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.380908 4894 scope.go:117] "RemoveContainer" containerID="29a7cf7f06bba0ab25f594e88e1ef9d5447382e4a77c6a0f4b2018f2477314c6" Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.382162 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nx7f7"] Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.391118 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nx7f7"] Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.419930 4894 scope.go:117] "RemoveContainer" containerID="a5ab5de61c62c1417a8aff6ce784c0af24c2a8908a0b2a526948c3a6fbc7c283" Dec 10 19:30:17 crc kubenswrapper[4894]: I1210 19:30:17.557560 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" path="/var/lib/kubelet/pods/de46bbe8-3eca-402c-b0ec-d41f4a70a469/volumes" Dec 10 19:30:19 crc kubenswrapper[4894]: I1210 19:30:19.576601 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:30:19 crc kubenswrapper[4894]: I1210 19:30:19.577176 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:30:22 crc kubenswrapper[4894]: I1210 19:30:22.401777 4894 generic.go:334] "Generic (PLEG): container finished" podID="bc921db2-6d84-4c34-9de5-12966910f500" containerID="2cf222abd7bf400cae8f70ebded8db2afb966af4e5336da3b23d321a33bc7c02" exitCode=0 Dec 10 19:30:22 crc kubenswrapper[4894]: I1210 19:30:22.402177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" event={"ID":"bc921db2-6d84-4c34-9de5-12966910f500","Type":"ContainerDied","Data":"2cf222abd7bf400cae8f70ebded8db2afb966af4e5336da3b23d321a33bc7c02"} Dec 10 19:30:22 crc kubenswrapper[4894]: I1210 19:30:22.656458 4894 scope.go:117] "RemoveContainer" containerID="8976b2ff473c23acd67f8a5a995c7cc92487d6fc62a56378d4562438b200f239" Dec 10 19:30:22 crc kubenswrapper[4894]: I1210 19:30:22.707578 4894 scope.go:117] "RemoveContainer" containerID="1b8571e70614c83cfc2f4179a1b158b448476733f2e469dce6ecac61f18f41fd" Dec 10 19:30:23 crc kubenswrapper[4894]: I1210 19:30:23.894484 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.008387 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-inventory\") pod \"bc921db2-6d84-4c34-9de5-12966910f500\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.008505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-ssh-key\") pod \"bc921db2-6d84-4c34-9de5-12966910f500\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.008599 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxp85\" (UniqueName: \"kubernetes.io/projected/bc921db2-6d84-4c34-9de5-12966910f500-kube-api-access-dxp85\") pod \"bc921db2-6d84-4c34-9de5-12966910f500\" (UID: \"bc921db2-6d84-4c34-9de5-12966910f500\") " Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.014569 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc921db2-6d84-4c34-9de5-12966910f500-kube-api-access-dxp85" (OuterVolumeSpecName: "kube-api-access-dxp85") pod "bc921db2-6d84-4c34-9de5-12966910f500" (UID: "bc921db2-6d84-4c34-9de5-12966910f500"). InnerVolumeSpecName "kube-api-access-dxp85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.043918 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-inventory" (OuterVolumeSpecName: "inventory") pod "bc921db2-6d84-4c34-9de5-12966910f500" (UID: "bc921db2-6d84-4c34-9de5-12966910f500"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.054028 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc921db2-6d84-4c34-9de5-12966910f500" (UID: "bc921db2-6d84-4c34-9de5-12966910f500"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.111161 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.111198 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxp85\" (UniqueName: \"kubernetes.io/projected/bc921db2-6d84-4c34-9de5-12966910f500-kube-api-access-dxp85\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.111212 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc921db2-6d84-4c34-9de5-12966910f500-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.426425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" event={"ID":"bc921db2-6d84-4c34-9de5-12966910f500","Type":"ContainerDied","Data":"1ff27bc5ac4f56d2951af68e63863b4624366a94cc3338226b9db4461cd08720"} Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.426467 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ff27bc5ac4f56d2951af68e63863b4624366a94cc3338226b9db4461cd08720" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.426478 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-597ft" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.523309 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-st6mh"] Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.524046 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc921db2-6d84-4c34-9de5-12966910f500" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.524172 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc921db2-6d84-4c34-9de5-12966910f500" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.524268 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="extract-content" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.524377 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="extract-content" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.524460 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="registry-server" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.524533 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="registry-server" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.524614 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6bdce4b-fa97-46a4-9293-b314220e0ebe" containerName="collect-profiles" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.524689 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6bdce4b-fa97-46a4-9293-b314220e0ebe" containerName="collect-profiles" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.524773 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="extract-utilities" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.524869 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="extract-utilities" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.524983 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="extract-utilities" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.525066 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="extract-utilities" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.525158 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="registry-server" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.525238 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="registry-server" Dec 10 19:30:24 crc kubenswrapper[4894]: E1210 19:30:24.525320 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="extract-content" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.525395 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="extract-content" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.525800 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f4c78a-52c7-48c1-be8c-0fbfa4859392" containerName="registry-server" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.525939 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="de46bbe8-3eca-402c-b0ec-d41f4a70a469" containerName="registry-server" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.526022 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6bdce4b-fa97-46a4-9293-b314220e0ebe" containerName="collect-profiles" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.526102 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc921db2-6d84-4c34-9de5-12966910f500" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.527190 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.529468 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.529638 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.529881 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.533839 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.534061 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-st6mh"] Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.620180 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.620329 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.620415 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgpzz\" (UniqueName: \"kubernetes.io/projected/94091bd6-87ce-46be-8410-f881ac9f0d66-kube-api-access-tgpzz\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.722785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.723266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.723373 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgpzz\" (UniqueName: \"kubernetes.io/projected/94091bd6-87ce-46be-8410-f881ac9f0d66-kube-api-access-tgpzz\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.726018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.729611 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.762716 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgpzz\" (UniqueName: \"kubernetes.io/projected/94091bd6-87ce-46be-8410-f881ac9f0d66-kube-api-access-tgpzz\") pod \"ssh-known-hosts-edpm-deployment-st6mh\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:24 crc kubenswrapper[4894]: I1210 19:30:24.846468 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:25 crc kubenswrapper[4894]: I1210 19:30:25.430386 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-st6mh"] Dec 10 19:30:25 crc kubenswrapper[4894]: I1210 19:30:25.431391 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:30:25 crc kubenswrapper[4894]: I1210 19:30:25.439919 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" event={"ID":"94091bd6-87ce-46be-8410-f881ac9f0d66","Type":"ContainerStarted","Data":"604657ace4b2fbe8c90670bd0507215956468b6d2256e28854700d7eb880e2d6"} Dec 10 19:30:26 crc kubenswrapper[4894]: I1210 19:30:26.451745 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" event={"ID":"94091bd6-87ce-46be-8410-f881ac9f0d66","Type":"ContainerStarted","Data":"ad260a698834535309ba15486f365263d732220fa0bd9886fbe0198af061af1b"} Dec 10 19:30:26 crc kubenswrapper[4894]: I1210 19:30:26.472016 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" podStartSLOduration=1.993727244 podStartE2EDuration="2.472000217s" podCreationTimestamp="2025-12-10 19:30:24 +0000 UTC" firstStartedPulling="2025-12-10 19:30:25.431130275 +0000 UTC m=+2108.225978067" lastFinishedPulling="2025-12-10 19:30:25.909403248 +0000 UTC m=+2108.704251040" observedRunningTime="2025-12-10 19:30:26.467855123 +0000 UTC m=+2109.262702915" watchObservedRunningTime="2025-12-10 19:30:26.472000217 +0000 UTC m=+2109.266848009" Dec 10 19:30:29 crc kubenswrapper[4894]: I1210 19:30:29.044031 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-zc7gw"] Dec 10 19:30:29 crc kubenswrapper[4894]: I1210 19:30:29.054615 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-zc7gw"] Dec 10 19:30:29 crc kubenswrapper[4894]: I1210 19:30:29.556502 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f972529a-80b2-4d4c-a9ed-1616052d506e" path="/var/lib/kubelet/pods/f972529a-80b2-4d4c-a9ed-1616052d506e/volumes" Dec 10 19:30:33 crc kubenswrapper[4894]: I1210 19:30:33.526361 4894 generic.go:334] "Generic (PLEG): container finished" podID="94091bd6-87ce-46be-8410-f881ac9f0d66" containerID="ad260a698834535309ba15486f365263d732220fa0bd9886fbe0198af061af1b" exitCode=0 Dec 10 19:30:33 crc kubenswrapper[4894]: I1210 19:30:33.526727 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" event={"ID":"94091bd6-87ce-46be-8410-f881ac9f0d66","Type":"ContainerDied","Data":"ad260a698834535309ba15486f365263d732220fa0bd9886fbe0198af061af1b"} Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.046086 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-8r6r4"] Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.066797 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-8r6r4"] Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.570774 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81df0908-03a5-439b-858d-e6e78ca7cce1" path="/var/lib/kubelet/pods/81df0908-03a5-439b-858d-e6e78ca7cce1/volumes" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.755582 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.873050 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-inventory-0\") pod \"94091bd6-87ce-46be-8410-f881ac9f0d66\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.873258 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgpzz\" (UniqueName: \"kubernetes.io/projected/94091bd6-87ce-46be-8410-f881ac9f0d66-kube-api-access-tgpzz\") pod \"94091bd6-87ce-46be-8410-f881ac9f0d66\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.873391 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-ssh-key-openstack-edpm-ipam\") pod \"94091bd6-87ce-46be-8410-f881ac9f0d66\" (UID: \"94091bd6-87ce-46be-8410-f881ac9f0d66\") " Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.878598 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94091bd6-87ce-46be-8410-f881ac9f0d66-kube-api-access-tgpzz" (OuterVolumeSpecName: "kube-api-access-tgpzz") pod "94091bd6-87ce-46be-8410-f881ac9f0d66" (UID: "94091bd6-87ce-46be-8410-f881ac9f0d66"). InnerVolumeSpecName "kube-api-access-tgpzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.905938 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "94091bd6-87ce-46be-8410-f881ac9f0d66" (UID: "94091bd6-87ce-46be-8410-f881ac9f0d66"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.915445 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "94091bd6-87ce-46be-8410-f881ac9f0d66" (UID: "94091bd6-87ce-46be-8410-f881ac9f0d66"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.975826 4894 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.975898 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgpzz\" (UniqueName: \"kubernetes.io/projected/94091bd6-87ce-46be-8410-f881ac9f0d66-kube-api-access-tgpzz\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:35 crc kubenswrapper[4894]: I1210 19:30:35.975910 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94091bd6-87ce-46be-8410-f881ac9f0d66-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.558519 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" event={"ID":"94091bd6-87ce-46be-8410-f881ac9f0d66","Type":"ContainerDied","Data":"604657ace4b2fbe8c90670bd0507215956468b6d2256e28854700d7eb880e2d6"} Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.558757 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="604657ace4b2fbe8c90670bd0507215956468b6d2256e28854700d7eb880e2d6" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.558814 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-st6mh" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.859474 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8"] Dec 10 19:30:36 crc kubenswrapper[4894]: E1210 19:30:36.860246 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94091bd6-87ce-46be-8410-f881ac9f0d66" containerName="ssh-known-hosts-edpm-deployment" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.860370 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="94091bd6-87ce-46be-8410-f881ac9f0d66" containerName="ssh-known-hosts-edpm-deployment" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.860788 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="94091bd6-87ce-46be-8410-f881ac9f0d66" containerName="ssh-known-hosts-edpm-deployment" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.861866 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.865712 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.865801 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.866570 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.866767 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.871701 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8"] Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.996745 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g2qj\" (UniqueName: \"kubernetes.io/projected/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-kube-api-access-8g2qj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.996889 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:36 crc kubenswrapper[4894]: I1210 19:30:36.996921 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.098190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g2qj\" (UniqueName: \"kubernetes.io/projected/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-kube-api-access-8g2qj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.098346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.098387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.103931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.105341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.114457 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g2qj\" (UniqueName: \"kubernetes.io/projected/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-kube-api-access-8g2qj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d4gh8\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.183592 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:37 crc kubenswrapper[4894]: I1210 19:30:37.720212 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8"] Dec 10 19:30:38 crc kubenswrapper[4894]: I1210 19:30:38.582753 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" event={"ID":"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0","Type":"ContainerStarted","Data":"2f5ef44c37b1a988944ac244223b4a2d26faf41be1ac77f5a9e130bf36557c34"} Dec 10 19:30:39 crc kubenswrapper[4894]: I1210 19:30:39.602452 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" event={"ID":"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0","Type":"ContainerStarted","Data":"edfc38c02d83aba719f05884caf85ef88693efa7f906fbece59573991f0222b7"} Dec 10 19:30:39 crc kubenswrapper[4894]: I1210 19:30:39.633963 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" podStartSLOduration=3.021029065 podStartE2EDuration="3.633908959s" podCreationTimestamp="2025-12-10 19:30:36 +0000 UTC" firstStartedPulling="2025-12-10 19:30:37.720401099 +0000 UTC m=+2120.515248891" lastFinishedPulling="2025-12-10 19:30:38.333280993 +0000 UTC m=+2121.128128785" observedRunningTime="2025-12-10 19:30:39.621584222 +0000 UTC m=+2122.416432024" watchObservedRunningTime="2025-12-10 19:30:39.633908959 +0000 UTC m=+2122.428756781" Dec 10 19:30:47 crc kubenswrapper[4894]: I1210 19:30:47.682146 4894 generic.go:334] "Generic (PLEG): container finished" podID="11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" containerID="edfc38c02d83aba719f05884caf85ef88693efa7f906fbece59573991f0222b7" exitCode=0 Dec 10 19:30:47 crc kubenswrapper[4894]: I1210 19:30:47.682227 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" event={"ID":"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0","Type":"ContainerDied","Data":"edfc38c02d83aba719f05884caf85ef88693efa7f906fbece59573991f0222b7"} Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.315410 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.486146 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g2qj\" (UniqueName: \"kubernetes.io/projected/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-kube-api-access-8g2qj\") pod \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.486213 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-ssh-key\") pod \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.486282 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-inventory\") pod \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\" (UID: \"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0\") " Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.519390 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-inventory" (OuterVolumeSpecName: "inventory") pod "11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" (UID: "11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.537023 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" (UID: "11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.576484 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.576535 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.588346 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.588385 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.599718 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-kube-api-access-8g2qj" (OuterVolumeSpecName: "kube-api-access-8g2qj") pod "11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" (UID: "11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0"). InnerVolumeSpecName "kube-api-access-8g2qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.690225 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g2qj\" (UniqueName: \"kubernetes.io/projected/11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0-kube-api-access-8g2qj\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.694190 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.698524 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a2a45ddb792c0c33a3e6ec306dbe31ee0324d8b7af5e3d8fe9603dedc9d4533"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.698606 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://1a2a45ddb792c0c33a3e6ec306dbe31ee0324d8b7af5e3d8fe9603dedc9d4533" gracePeriod=600 Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.705232 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" event={"ID":"11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0","Type":"ContainerDied","Data":"2f5ef44c37b1a988944ac244223b4a2d26faf41be1ac77f5a9e130bf36557c34"} Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.705277 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f5ef44c37b1a988944ac244223b4a2d26faf41be1ac77f5a9e130bf36557c34" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.705350 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d4gh8" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.777046 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n"] Dec 10 19:30:49 crc kubenswrapper[4894]: E1210 19:30:49.777701 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.777744 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.778071 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.778887 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.780669 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.781133 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.781156 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.781836 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.813019 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n"] Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.904624 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbfz\" (UniqueName: \"kubernetes.io/projected/710e20ed-7ffe-4041-9220-8ffa6b12b565-kube-api-access-vzbfz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.904801 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:49 crc kubenswrapper[4894]: I1210 19:30:49.904864 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.006823 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.007218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.007302 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbfz\" (UniqueName: \"kubernetes.io/projected/710e20ed-7ffe-4041-9220-8ffa6b12b565-kube-api-access-vzbfz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.013486 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.013538 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.029622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbfz\" (UniqueName: \"kubernetes.io/projected/710e20ed-7ffe-4041-9220-8ffa6b12b565-kube-api-access-vzbfz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.106967 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.717950 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="1a2a45ddb792c0c33a3e6ec306dbe31ee0324d8b7af5e3d8fe9603dedc9d4533" exitCode=0 Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.718110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"1a2a45ddb792c0c33a3e6ec306dbe31ee0324d8b7af5e3d8fe9603dedc9d4533"} Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.718404 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a"} Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.718443 4894 scope.go:117] "RemoveContainer" containerID="10a39afa9a0535dcd126d68cb82dff16fd486e4bc224e98c4c4d689539c93eb4" Dec 10 19:30:50 crc kubenswrapper[4894]: W1210 19:30:50.730113 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod710e20ed_7ffe_4041_9220_8ffa6b12b565.slice/crio-5b27cb6acf9b5d7a4bd5ccfe6430b5b177a6944de2bd68950fc3afd32e52edd6 WatchSource:0}: Error finding container 5b27cb6acf9b5d7a4bd5ccfe6430b5b177a6944de2bd68950fc3afd32e52edd6: Status 404 returned error can't find the container with id 5b27cb6acf9b5d7a4bd5ccfe6430b5b177a6944de2bd68950fc3afd32e52edd6 Dec 10 19:30:50 crc kubenswrapper[4894]: I1210 19:30:50.738613 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n"] Dec 10 19:30:51 crc kubenswrapper[4894]: I1210 19:30:51.740055 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" event={"ID":"710e20ed-7ffe-4041-9220-8ffa6b12b565","Type":"ContainerStarted","Data":"1f982348aaec7f0f3035b47e9523704411f7e636815c42b5d4be55c74ff7ffd5"} Dec 10 19:30:51 crc kubenswrapper[4894]: I1210 19:30:51.740399 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" event={"ID":"710e20ed-7ffe-4041-9220-8ffa6b12b565","Type":"ContainerStarted","Data":"5b27cb6acf9b5d7a4bd5ccfe6430b5b177a6944de2bd68950fc3afd32e52edd6"} Dec 10 19:30:51 crc kubenswrapper[4894]: I1210 19:30:51.762545 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" podStartSLOduration=2.29459553 podStartE2EDuration="2.762529069s" podCreationTimestamp="2025-12-10 19:30:49 +0000 UTC" firstStartedPulling="2025-12-10 19:30:50.733002009 +0000 UTC m=+2133.527849801" lastFinishedPulling="2025-12-10 19:30:51.200935548 +0000 UTC m=+2133.995783340" observedRunningTime="2025-12-10 19:30:51.752500254 +0000 UTC m=+2134.547348056" watchObservedRunningTime="2025-12-10 19:30:51.762529069 +0000 UTC m=+2134.557376861" Dec 10 19:31:01 crc kubenswrapper[4894]: I1210 19:31:01.853573 4894 generic.go:334] "Generic (PLEG): container finished" podID="710e20ed-7ffe-4041-9220-8ffa6b12b565" containerID="1f982348aaec7f0f3035b47e9523704411f7e636815c42b5d4be55c74ff7ffd5" exitCode=0 Dec 10 19:31:01 crc kubenswrapper[4894]: I1210 19:31:01.853676 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" event={"ID":"710e20ed-7ffe-4041-9220-8ffa6b12b565","Type":"ContainerDied","Data":"1f982348aaec7f0f3035b47e9523704411f7e636815c42b5d4be55c74ff7ffd5"} Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.396679 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.505371 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-ssh-key\") pod \"710e20ed-7ffe-4041-9220-8ffa6b12b565\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.506008 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-inventory\") pod \"710e20ed-7ffe-4041-9220-8ffa6b12b565\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.506153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzbfz\" (UniqueName: \"kubernetes.io/projected/710e20ed-7ffe-4041-9220-8ffa6b12b565-kube-api-access-vzbfz\") pod \"710e20ed-7ffe-4041-9220-8ffa6b12b565\" (UID: \"710e20ed-7ffe-4041-9220-8ffa6b12b565\") " Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.513997 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/710e20ed-7ffe-4041-9220-8ffa6b12b565-kube-api-access-vzbfz" (OuterVolumeSpecName: "kube-api-access-vzbfz") pod "710e20ed-7ffe-4041-9220-8ffa6b12b565" (UID: "710e20ed-7ffe-4041-9220-8ffa6b12b565"). InnerVolumeSpecName "kube-api-access-vzbfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.548117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "710e20ed-7ffe-4041-9220-8ffa6b12b565" (UID: "710e20ed-7ffe-4041-9220-8ffa6b12b565"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.561624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-inventory" (OuterVolumeSpecName: "inventory") pod "710e20ed-7ffe-4041-9220-8ffa6b12b565" (UID: "710e20ed-7ffe-4041-9220-8ffa6b12b565"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.609178 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.609222 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzbfz\" (UniqueName: \"kubernetes.io/projected/710e20ed-7ffe-4041-9220-8ffa6b12b565-kube-api-access-vzbfz\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.609243 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710e20ed-7ffe-4041-9220-8ffa6b12b565-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.877291 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" event={"ID":"710e20ed-7ffe-4041-9220-8ffa6b12b565","Type":"ContainerDied","Data":"5b27cb6acf9b5d7a4bd5ccfe6430b5b177a6944de2bd68950fc3afd32e52edd6"} Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.877329 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b27cb6acf9b5d7a4bd5ccfe6430b5b177a6944de2bd68950fc3afd32e52edd6" Dec 10 19:31:03 crc kubenswrapper[4894]: I1210 19:31:03.877397 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.036779 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh"] Dec 10 19:31:04 crc kubenswrapper[4894]: E1210 19:31:04.037943 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710e20ed-7ffe-4041-9220-8ffa6b12b565" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.038087 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="710e20ed-7ffe-4041-9220-8ffa6b12b565" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.038528 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="710e20ed-7ffe-4041-9220-8ffa6b12b565" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.039927 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.044174 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.044216 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.044352 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.044521 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.044924 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.044936 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.045205 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.045281 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.045407 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh"] Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223079 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223251 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223303 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223435 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223470 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223497 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223571 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223609 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.223930 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62ng5\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-kube-api-access-62ng5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.224083 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.224161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.326255 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.326369 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.326473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.326521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62ng5\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-kube-api-access-62ng5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.326614 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.327627 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.327818 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.327958 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.328068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.328124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.328206 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.328310 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.328385 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.328446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.331868 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.332828 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.333605 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.333818 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.334711 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.337955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.343084 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.343330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.344410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.344969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.344967 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.348534 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.349966 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.356112 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62ng5\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-kube-api-access-62ng5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-556fh\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.374644 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:04 crc kubenswrapper[4894]: I1210 19:31:04.975839 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh"] Dec 10 19:31:05 crc kubenswrapper[4894]: I1210 19:31:05.897889 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" event={"ID":"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb","Type":"ContainerStarted","Data":"f00729175d8002012e1757598a4fe43f09559253c877d614085afb11e2aaac0d"} Dec 10 19:31:06 crc kubenswrapper[4894]: I1210 19:31:06.913208 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" event={"ID":"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb","Type":"ContainerStarted","Data":"73db0edb8107230e9c4c1218d03927996702ec4d1611d4a2020a8262d2e9b476"} Dec 10 19:31:06 crc kubenswrapper[4894]: I1210 19:31:06.934222 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" podStartSLOduration=2.206923631 podStartE2EDuration="2.934204427s" podCreationTimestamp="2025-12-10 19:31:04 +0000 UTC" firstStartedPulling="2025-12-10 19:31:04.983597858 +0000 UTC m=+2147.778445650" lastFinishedPulling="2025-12-10 19:31:05.710878614 +0000 UTC m=+2148.505726446" observedRunningTime="2025-12-10 19:31:06.93064514 +0000 UTC m=+2149.725492952" watchObservedRunningTime="2025-12-10 19:31:06.934204427 +0000 UTC m=+2149.729052229" Dec 10 19:31:22 crc kubenswrapper[4894]: I1210 19:31:22.857832 4894 scope.go:117] "RemoveContainer" containerID="a699da97fb3212b780700797a9bdc9ff5a5855c7df869e685cfe7b2046b818a9" Dec 10 19:31:22 crc kubenswrapper[4894]: I1210 19:31:22.898569 4894 scope.go:117] "RemoveContainer" containerID="ade95ffd9ee9fe93a57925519398a4c08dd6ac8215306564fad730aff70cf28e" Dec 10 19:31:48 crc kubenswrapper[4894]: I1210 19:31:48.337616 4894 generic.go:334] "Generic (PLEG): container finished" podID="1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" containerID="73db0edb8107230e9c4c1218d03927996702ec4d1611d4a2020a8262d2e9b476" exitCode=0 Dec 10 19:31:48 crc kubenswrapper[4894]: I1210 19:31:48.337732 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" event={"ID":"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb","Type":"ContainerDied","Data":"73db0edb8107230e9c4c1218d03927996702ec4d1611d4a2020a8262d2e9b476"} Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.027469 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163075 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-bootstrap-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-nova-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163206 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62ng5\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-kube-api-access-62ng5\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163232 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-libvirt-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163314 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ssh-key\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163349 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163440 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163485 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ovn-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163530 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163563 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-inventory\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163639 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-ovn-default-certs-0\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163683 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-neutron-metadata-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163713 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-telemetry-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.163759 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-repo-setup-combined-ca-bundle\") pod \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\" (UID: \"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb\") " Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.170668 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.170708 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-kube-api-access-62ng5" (OuterVolumeSpecName: "kube-api-access-62ng5") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "kube-api-access-62ng5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.173092 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.173119 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.173492 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.173760 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.174048 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.175394 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.176411 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.177008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.177236 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.177578 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.200578 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-inventory" (OuterVolumeSpecName: "inventory") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.218521 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" (UID: "1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267264 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267329 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267353 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267374 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267392 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267414 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267432 4894 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267450 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267467 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267484 4894 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267501 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62ng5\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-kube-api-access-62ng5\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267520 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267537 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.267560 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.363182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" event={"ID":"1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb","Type":"ContainerDied","Data":"f00729175d8002012e1757598a4fe43f09559253c877d614085afb11e2aaac0d"} Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.363241 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f00729175d8002012e1757598a4fe43f09559253c877d614085afb11e2aaac0d" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.363313 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-556fh" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.498951 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7"] Dec 10 19:31:50 crc kubenswrapper[4894]: E1210 19:31:50.499447 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.499467 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.499658 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.500505 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.502907 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.503060 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.503765 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.504077 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.504523 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.518381 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7"] Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.580174 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.580385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.580623 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.580758 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.580828 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx7bk\" (UniqueName: \"kubernetes.io/projected/c91b533a-4f71-4b78-9442-a6d7bf300bff-kube-api-access-gx7bk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.682997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.683061 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.683134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.683171 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.683234 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx7bk\" (UniqueName: \"kubernetes.io/projected/c91b533a-4f71-4b78-9442-a6d7bf300bff-kube-api-access-gx7bk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.684573 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.686534 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.687118 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.688136 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.704202 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx7bk\" (UniqueName: \"kubernetes.io/projected/c91b533a-4f71-4b78-9442-a6d7bf300bff-kube-api-access-gx7bk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fhwr7\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:50 crc kubenswrapper[4894]: I1210 19:31:50.836074 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:31:51 crc kubenswrapper[4894]: I1210 19:31:51.461734 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7"] Dec 10 19:31:52 crc kubenswrapper[4894]: I1210 19:31:52.391314 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" event={"ID":"c91b533a-4f71-4b78-9442-a6d7bf300bff","Type":"ContainerStarted","Data":"1f6fb3888e30bc0b0e6a408830ea0fc3e9ce0af3239ad4b2608fc5b7e3f66050"} Dec 10 19:31:52 crc kubenswrapper[4894]: I1210 19:31:52.391725 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" event={"ID":"c91b533a-4f71-4b78-9442-a6d7bf300bff","Type":"ContainerStarted","Data":"e613e78b3055781e9f9ef08697b2107633ac14383eda4289e9ede1a9528334d1"} Dec 10 19:31:52 crc kubenswrapper[4894]: I1210 19:31:52.422330 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" podStartSLOduration=1.9450405370000001 podStartE2EDuration="2.422302234s" podCreationTimestamp="2025-12-10 19:31:50 +0000 UTC" firstStartedPulling="2025-12-10 19:31:51.463759311 +0000 UTC m=+2194.258607143" lastFinishedPulling="2025-12-10 19:31:51.941021038 +0000 UTC m=+2194.735868840" observedRunningTime="2025-12-10 19:31:52.410376499 +0000 UTC m=+2195.205224301" watchObservedRunningTime="2025-12-10 19:31:52.422302234 +0000 UTC m=+2195.217150056" Dec 10 19:32:49 crc kubenswrapper[4894]: I1210 19:32:49.577058 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:32:49 crc kubenswrapper[4894]: I1210 19:32:49.577816 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:33:02 crc kubenswrapper[4894]: I1210 19:33:02.165881 4894 generic.go:334] "Generic (PLEG): container finished" podID="c91b533a-4f71-4b78-9442-a6d7bf300bff" containerID="1f6fb3888e30bc0b0e6a408830ea0fc3e9ce0af3239ad4b2608fc5b7e3f66050" exitCode=0 Dec 10 19:33:02 crc kubenswrapper[4894]: I1210 19:33:02.165995 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" event={"ID":"c91b533a-4f71-4b78-9442-a6d7bf300bff","Type":"ContainerDied","Data":"1f6fb3888e30bc0b0e6a408830ea0fc3e9ce0af3239ad4b2608fc5b7e3f66050"} Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.232178 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7frm4"] Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.246458 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.258572 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7frm4"] Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.439026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-utilities\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.439154 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nj9n\" (UniqueName: \"kubernetes.io/projected/a0e478af-801f-4f79-a129-d55974fdc1c1-kube-api-access-8nj9n\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.439180 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-catalog-content\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.541552 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nj9n\" (UniqueName: \"kubernetes.io/projected/a0e478af-801f-4f79-a129-d55974fdc1c1-kube-api-access-8nj9n\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.541611 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-catalog-content\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.541768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-utilities\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.542511 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-utilities\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.542535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-catalog-content\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.561035 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nj9n\" (UniqueName: \"kubernetes.io/projected/a0e478af-801f-4f79-a129-d55974fdc1c1-kube-api-access-8nj9n\") pod \"certified-operators-7frm4\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.581347 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.768083 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.955760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ssh-key\") pod \"c91b533a-4f71-4b78-9442-a6d7bf300bff\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.955829 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-inventory\") pod \"c91b533a-4f71-4b78-9442-a6d7bf300bff\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.955875 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovn-combined-ca-bundle\") pod \"c91b533a-4f71-4b78-9442-a6d7bf300bff\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.956014 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx7bk\" (UniqueName: \"kubernetes.io/projected/c91b533a-4f71-4b78-9442-a6d7bf300bff-kube-api-access-gx7bk\") pod \"c91b533a-4f71-4b78-9442-a6d7bf300bff\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.956060 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovncontroller-config-0\") pod \"c91b533a-4f71-4b78-9442-a6d7bf300bff\" (UID: \"c91b533a-4f71-4b78-9442-a6d7bf300bff\") " Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.964301 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91b533a-4f71-4b78-9442-a6d7bf300bff-kube-api-access-gx7bk" (OuterVolumeSpecName: "kube-api-access-gx7bk") pod "c91b533a-4f71-4b78-9442-a6d7bf300bff" (UID: "c91b533a-4f71-4b78-9442-a6d7bf300bff"). InnerVolumeSpecName "kube-api-access-gx7bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:33:03 crc kubenswrapper[4894]: I1210 19:33:03.977409 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c91b533a-4f71-4b78-9442-a6d7bf300bff" (UID: "c91b533a-4f71-4b78-9442-a6d7bf300bff"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.010142 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c91b533a-4f71-4b78-9442-a6d7bf300bff" (UID: "c91b533a-4f71-4b78-9442-a6d7bf300bff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.026031 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-inventory" (OuterVolumeSpecName: "inventory") pod "c91b533a-4f71-4b78-9442-a6d7bf300bff" (UID: "c91b533a-4f71-4b78-9442-a6d7bf300bff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.043444 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c91b533a-4f71-4b78-9442-a6d7bf300bff" (UID: "c91b533a-4f71-4b78-9442-a6d7bf300bff"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.058134 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.058172 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.058185 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.058202 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx7bk\" (UniqueName: \"kubernetes.io/projected/c91b533a-4f71-4b78-9442-a6d7bf300bff-kube-api-access-gx7bk\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.058213 4894 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c91b533a-4f71-4b78-9442-a6d7bf300bff-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.188816 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" event={"ID":"c91b533a-4f71-4b78-9442-a6d7bf300bff","Type":"ContainerDied","Data":"e613e78b3055781e9f9ef08697b2107633ac14383eda4289e9ede1a9528334d1"} Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.188877 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e613e78b3055781e9f9ef08697b2107633ac14383eda4289e9ede1a9528334d1" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.188887 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fhwr7" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.264492 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7frm4"] Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.296340 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg"] Dec 10 19:33:04 crc kubenswrapper[4894]: E1210 19:33:04.320756 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91b533a-4f71-4b78-9442-a6d7bf300bff" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.320788 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91b533a-4f71-4b78-9442-a6d7bf300bff" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.321398 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91b533a-4f71-4b78-9442-a6d7bf300bff" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.322533 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.328470 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.328763 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.328920 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.329749 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.329977 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.339968 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.356891 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg"] Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.471452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.471502 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.471545 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.471641 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.471684 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8rgv\" (UniqueName: \"kubernetes.io/projected/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-kube-api-access-t8rgv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.471712 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.573993 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8rgv\" (UniqueName: \"kubernetes.io/projected/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-kube-api-access-t8rgv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.574609 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.575418 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.575781 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.575912 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.576059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.586501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.586545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.587320 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.589633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.590330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.597824 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8rgv\" (UniqueName: \"kubernetes.io/projected/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-kube-api-access-t8rgv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:04 crc kubenswrapper[4894]: I1210 19:33:04.655361 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:05 crc kubenswrapper[4894]: I1210 19:33:05.208684 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg"] Dec 10 19:33:05 crc kubenswrapper[4894]: I1210 19:33:05.209709 4894 generic.go:334] "Generic (PLEG): container finished" podID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerID="eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25" exitCode=0 Dec 10 19:33:05 crc kubenswrapper[4894]: I1210 19:33:05.209741 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerDied","Data":"eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25"} Dec 10 19:33:05 crc kubenswrapper[4894]: I1210 19:33:05.209758 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerStarted","Data":"99642d896b3a06753ce61cd1a884a46e6fa46239086a012e0e20a86f5feb55a8"} Dec 10 19:33:05 crc kubenswrapper[4894]: W1210 19:33:05.223466 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9368153_c2ba_42ac_b7a5_5d47b02d8d14.slice/crio-2a034fde7e569fbd30bc4f31483be814a3b160d28769506924c311d308ad5197 WatchSource:0}: Error finding container 2a034fde7e569fbd30bc4f31483be814a3b160d28769506924c311d308ad5197: Status 404 returned error can't find the container with id 2a034fde7e569fbd30bc4f31483be814a3b160d28769506924c311d308ad5197 Dec 10 19:33:06 crc kubenswrapper[4894]: I1210 19:33:06.220368 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerStarted","Data":"34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f"} Dec 10 19:33:06 crc kubenswrapper[4894]: I1210 19:33:06.222377 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" event={"ID":"a9368153-c2ba-42ac-b7a5-5d47b02d8d14","Type":"ContainerStarted","Data":"4d71d499c361ea6b3a273fe76bdd0ef68b500db00a25341931e88b1cda496456"} Dec 10 19:33:06 crc kubenswrapper[4894]: I1210 19:33:06.222412 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" event={"ID":"a9368153-c2ba-42ac-b7a5-5d47b02d8d14","Type":"ContainerStarted","Data":"2a034fde7e569fbd30bc4f31483be814a3b160d28769506924c311d308ad5197"} Dec 10 19:33:06 crc kubenswrapper[4894]: I1210 19:33:06.264573 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" podStartSLOduration=1.730840027 podStartE2EDuration="2.264550212s" podCreationTimestamp="2025-12-10 19:33:04 +0000 UTC" firstStartedPulling="2025-12-10 19:33:05.227167161 +0000 UTC m=+2268.022014953" lastFinishedPulling="2025-12-10 19:33:05.760877336 +0000 UTC m=+2268.555725138" observedRunningTime="2025-12-10 19:33:06.255636379 +0000 UTC m=+2269.050484171" watchObservedRunningTime="2025-12-10 19:33:06.264550212 +0000 UTC m=+2269.059398014" Dec 10 19:33:07 crc kubenswrapper[4894]: I1210 19:33:07.441743 4894 generic.go:334] "Generic (PLEG): container finished" podID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerID="34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f" exitCode=0 Dec 10 19:33:07 crc kubenswrapper[4894]: I1210 19:33:07.441901 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerDied","Data":"34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f"} Dec 10 19:33:08 crc kubenswrapper[4894]: I1210 19:33:08.454782 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerStarted","Data":"2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58"} Dec 10 19:33:08 crc kubenswrapper[4894]: I1210 19:33:08.488458 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7frm4" podStartSLOduration=2.768245916 podStartE2EDuration="5.488439043s" podCreationTimestamp="2025-12-10 19:33:03 +0000 UTC" firstStartedPulling="2025-12-10 19:33:05.211775762 +0000 UTC m=+2268.006623594" lastFinishedPulling="2025-12-10 19:33:07.931968939 +0000 UTC m=+2270.726816721" observedRunningTime="2025-12-10 19:33:08.477705861 +0000 UTC m=+2271.272553653" watchObservedRunningTime="2025-12-10 19:33:08.488439043 +0000 UTC m=+2271.283286855" Dec 10 19:33:13 crc kubenswrapper[4894]: I1210 19:33:13.582048 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:13 crc kubenswrapper[4894]: I1210 19:33:13.582561 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:13 crc kubenswrapper[4894]: I1210 19:33:13.632006 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:14 crc kubenswrapper[4894]: I1210 19:33:14.588442 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:14 crc kubenswrapper[4894]: I1210 19:33:14.644998 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7frm4"] Dec 10 19:33:16 crc kubenswrapper[4894]: I1210 19:33:16.533274 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7frm4" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="registry-server" containerID="cri-o://2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58" gracePeriod=2 Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.120516 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.231378 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nj9n\" (UniqueName: \"kubernetes.io/projected/a0e478af-801f-4f79-a129-d55974fdc1c1-kube-api-access-8nj9n\") pod \"a0e478af-801f-4f79-a129-d55974fdc1c1\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.231578 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-utilities\") pod \"a0e478af-801f-4f79-a129-d55974fdc1c1\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.231614 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-catalog-content\") pod \"a0e478af-801f-4f79-a129-d55974fdc1c1\" (UID: \"a0e478af-801f-4f79-a129-d55974fdc1c1\") " Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.232491 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-utilities" (OuterVolumeSpecName: "utilities") pod "a0e478af-801f-4f79-a129-d55974fdc1c1" (UID: "a0e478af-801f-4f79-a129-d55974fdc1c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.237415 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0e478af-801f-4f79-a129-d55974fdc1c1-kube-api-access-8nj9n" (OuterVolumeSpecName: "kube-api-access-8nj9n") pod "a0e478af-801f-4f79-a129-d55974fdc1c1" (UID: "a0e478af-801f-4f79-a129-d55974fdc1c1"). InnerVolumeSpecName "kube-api-access-8nj9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.333642 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.333671 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nj9n\" (UniqueName: \"kubernetes.io/projected/a0e478af-801f-4f79-a129-d55974fdc1c1-kube-api-access-8nj9n\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.505668 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0e478af-801f-4f79-a129-d55974fdc1c1" (UID: "a0e478af-801f-4f79-a129-d55974fdc1c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.537428 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0e478af-801f-4f79-a129-d55974fdc1c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.551823 4894 generic.go:334] "Generic (PLEG): container finished" podID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerID="2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58" exitCode=0 Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.552184 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7frm4" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.556666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerDied","Data":"2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58"} Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.556700 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7frm4" event={"ID":"a0e478af-801f-4f79-a129-d55974fdc1c1","Type":"ContainerDied","Data":"99642d896b3a06753ce61cd1a884a46e6fa46239086a012e0e20a86f5feb55a8"} Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.556751 4894 scope.go:117] "RemoveContainer" containerID="2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.583150 4894 scope.go:117] "RemoveContainer" containerID="34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.604012 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7frm4"] Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.616407 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7frm4"] Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.619321 4894 scope.go:117] "RemoveContainer" containerID="eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.675802 4894 scope.go:117] "RemoveContainer" containerID="2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58" Dec 10 19:33:17 crc kubenswrapper[4894]: E1210 19:33:17.676656 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58\": container with ID starting with 2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58 not found: ID does not exist" containerID="2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.676714 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58"} err="failed to get container status \"2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58\": rpc error: code = NotFound desc = could not find container \"2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58\": container with ID starting with 2967050dcc65105d9f4f5542691603c939d178545888ba032ca054f215218b58 not found: ID does not exist" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.676746 4894 scope.go:117] "RemoveContainer" containerID="34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f" Dec 10 19:33:17 crc kubenswrapper[4894]: E1210 19:33:17.677252 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f\": container with ID starting with 34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f not found: ID does not exist" containerID="34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.677296 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f"} err="failed to get container status \"34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f\": rpc error: code = NotFound desc = could not find container \"34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f\": container with ID starting with 34dc8578e6c4e6f5c5422fd8a2e014b9b53b4962db6d10da29f7bbc12674eb2f not found: ID does not exist" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.677324 4894 scope.go:117] "RemoveContainer" containerID="eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25" Dec 10 19:33:17 crc kubenswrapper[4894]: E1210 19:33:17.677614 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25\": container with ID starting with eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25 not found: ID does not exist" containerID="eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25" Dec 10 19:33:17 crc kubenswrapper[4894]: I1210 19:33:17.677652 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25"} err="failed to get container status \"eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25\": rpc error: code = NotFound desc = could not find container \"eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25\": container with ID starting with eb7881509caf3d82d564ffca77f14735b9a9542f17e89430b909fc8e02567f25 not found: ID does not exist" Dec 10 19:33:19 crc kubenswrapper[4894]: I1210 19:33:19.565176 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" path="/var/lib/kubelet/pods/a0e478af-801f-4f79-a129-d55974fdc1c1/volumes" Dec 10 19:33:19 crc kubenswrapper[4894]: I1210 19:33:19.577075 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:33:19 crc kubenswrapper[4894]: I1210 19:33:19.577146 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.578232 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.578875 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.578945 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.579811 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.579895 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" gracePeriod=600 Dec 10 19:33:49 crc kubenswrapper[4894]: E1210 19:33:49.704673 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.896565 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" exitCode=0 Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.896625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a"} Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.896693 4894 scope.go:117] "RemoveContainer" containerID="1a2a45ddb792c0c33a3e6ec306dbe31ee0324d8b7af5e3d8fe9603dedc9d4533" Dec 10 19:33:49 crc kubenswrapper[4894]: I1210 19:33:49.897367 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:33:49 crc kubenswrapper[4894]: E1210 19:33:49.897660 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:33:58 crc kubenswrapper[4894]: I1210 19:33:58.012777 4894 generic.go:334] "Generic (PLEG): container finished" podID="a9368153-c2ba-42ac-b7a5-5d47b02d8d14" containerID="4d71d499c361ea6b3a273fe76bdd0ef68b500db00a25341931e88b1cda496456" exitCode=0 Dec 10 19:33:58 crc kubenswrapper[4894]: I1210 19:33:58.013493 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" event={"ID":"a9368153-c2ba-42ac-b7a5-5d47b02d8d14","Type":"ContainerDied","Data":"4d71d499c361ea6b3a273fe76bdd0ef68b500db00a25341931e88b1cda496456"} Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.569690 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.590169 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.590258 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-inventory\") pod \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.590284 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-nova-metadata-neutron-config-0\") pod \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.590304 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8rgv\" (UniqueName: \"kubernetes.io/projected/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-kube-api-access-t8rgv\") pod \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.590348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-metadata-combined-ca-bundle\") pod \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.590368 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-ssh-key\") pod \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\" (UID: \"a9368153-c2ba-42ac-b7a5-5d47b02d8d14\") " Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.602281 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a9368153-c2ba-42ac-b7a5-5d47b02d8d14" (UID: "a9368153-c2ba-42ac-b7a5-5d47b02d8d14"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.602613 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-kube-api-access-t8rgv" (OuterVolumeSpecName: "kube-api-access-t8rgv") pod "a9368153-c2ba-42ac-b7a5-5d47b02d8d14" (UID: "a9368153-c2ba-42ac-b7a5-5d47b02d8d14"). InnerVolumeSpecName "kube-api-access-t8rgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.647696 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a9368153-c2ba-42ac-b7a5-5d47b02d8d14" (UID: "a9368153-c2ba-42ac-b7a5-5d47b02d8d14"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.647724 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-inventory" (OuterVolumeSpecName: "inventory") pod "a9368153-c2ba-42ac-b7a5-5d47b02d8d14" (UID: "a9368153-c2ba-42ac-b7a5-5d47b02d8d14"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.651054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a9368153-c2ba-42ac-b7a5-5d47b02d8d14" (UID: "a9368153-c2ba-42ac-b7a5-5d47b02d8d14"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.658548 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a9368153-c2ba-42ac-b7a5-5d47b02d8d14" (UID: "a9368153-c2ba-42ac-b7a5-5d47b02d8d14"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.692788 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.692822 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.692837 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8rgv\" (UniqueName: \"kubernetes.io/projected/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-kube-api-access-t8rgv\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.692868 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.692880 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:59 crc kubenswrapper[4894]: I1210 19:33:59.692891 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a9368153-c2ba-42ac-b7a5-5d47b02d8d14-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.043587 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" event={"ID":"a9368153-c2ba-42ac-b7a5-5d47b02d8d14","Type":"ContainerDied","Data":"2a034fde7e569fbd30bc4f31483be814a3b160d28769506924c311d308ad5197"} Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.043634 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a034fde7e569fbd30bc4f31483be814a3b160d28769506924c311d308ad5197" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.043645 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.143291 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf"] Dec 10 19:34:00 crc kubenswrapper[4894]: E1210 19:34:00.143685 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="registry-server" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.143703 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="registry-server" Dec 10 19:34:00 crc kubenswrapper[4894]: E1210 19:34:00.143738 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="extract-utilities" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.143745 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="extract-utilities" Dec 10 19:34:00 crc kubenswrapper[4894]: E1210 19:34:00.143756 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="extract-content" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.143762 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="extract-content" Dec 10 19:34:00 crc kubenswrapper[4894]: E1210 19:34:00.143775 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9368153-c2ba-42ac-b7a5-5d47b02d8d14" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.143783 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9368153-c2ba-42ac-b7a5-5d47b02d8d14" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.144023 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9368153-c2ba-42ac-b7a5-5d47b02d8d14" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.144046 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e478af-801f-4f79-a129-d55974fdc1c1" containerName="registry-server" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.145100 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.147290 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.147559 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.147743 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.148072 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.148296 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.159308 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf"] Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.303865 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.303930 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.304006 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.304133 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dfrn\" (UniqueName: \"kubernetes.io/projected/a5f93c12-d3db-42db-b431-861c46535ba4-kube-api-access-7dfrn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.304241 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.407012 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dfrn\" (UniqueName: \"kubernetes.io/projected/a5f93c12-d3db-42db-b431-861c46535ba4-kube-api-access-7dfrn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.407106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.407232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.407266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.407317 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.412534 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.412581 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.413107 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.413260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.427724 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dfrn\" (UniqueName: \"kubernetes.io/projected/a5f93c12-d3db-42db-b431-861c46535ba4-kube-api-access-7dfrn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.465526 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:34:00 crc kubenswrapper[4894]: I1210 19:34:00.541691 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:34:00 crc kubenswrapper[4894]: E1210 19:34:00.541959 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:34:01 crc kubenswrapper[4894]: I1210 19:34:01.201968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf"] Dec 10 19:34:02 crc kubenswrapper[4894]: I1210 19:34:02.065560 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" event={"ID":"a5f93c12-d3db-42db-b431-861c46535ba4","Type":"ContainerStarted","Data":"e5cb2c8d6de2f799e87f0a9b94e84ef9e0e5a26a0572578e14bf7c1dabf6c850"} Dec 10 19:34:03 crc kubenswrapper[4894]: I1210 19:34:03.074890 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" event={"ID":"a5f93c12-d3db-42db-b431-861c46535ba4","Type":"ContainerStarted","Data":"3ca07f4eb9bf413167d007f63917d8065fea7281e2c4596b7c1f6a16be6aeafe"} Dec 10 19:34:03 crc kubenswrapper[4894]: I1210 19:34:03.111699 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" podStartSLOduration=2.4381824 podStartE2EDuration="3.111679481s" podCreationTimestamp="2025-12-10 19:34:00 +0000 UTC" firstStartedPulling="2025-12-10 19:34:01.205642145 +0000 UTC m=+2324.000489937" lastFinishedPulling="2025-12-10 19:34:01.879139226 +0000 UTC m=+2324.673987018" observedRunningTime="2025-12-10 19:34:03.10649213 +0000 UTC m=+2325.901339942" watchObservedRunningTime="2025-12-10 19:34:03.111679481 +0000 UTC m=+2325.906527273" Dec 10 19:34:13 crc kubenswrapper[4894]: I1210 19:34:13.541067 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:34:13 crc kubenswrapper[4894]: E1210 19:34:13.541880 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:34:28 crc kubenswrapper[4894]: I1210 19:34:28.541345 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:34:28 crc kubenswrapper[4894]: E1210 19:34:28.542276 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.327129 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j2ch5"] Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.330543 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.351442 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j2ch5"] Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.465279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-utilities\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.465751 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnzz2\" (UniqueName: \"kubernetes.io/projected/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-kube-api-access-jnzz2\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.465798 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-catalog-content\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.567498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-catalog-content\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.567644 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-utilities\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.567791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnzz2\" (UniqueName: \"kubernetes.io/projected/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-kube-api-access-jnzz2\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.568596 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-catalog-content\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.569037 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-utilities\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.595526 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnzz2\" (UniqueName: \"kubernetes.io/projected/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-kube-api-access-jnzz2\") pod \"community-operators-j2ch5\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:33 crc kubenswrapper[4894]: I1210 19:34:33.673752 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:34 crc kubenswrapper[4894]: I1210 19:34:34.238969 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j2ch5"] Dec 10 19:34:34 crc kubenswrapper[4894]: I1210 19:34:34.437616 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerStarted","Data":"a369e82e783fe82474cd5d993f24c6b61224f05ebfd82aafdd9ddecb22a90426"} Dec 10 19:34:35 crc kubenswrapper[4894]: I1210 19:34:35.453408 4894 generic.go:334] "Generic (PLEG): container finished" podID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerID="575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62" exitCode=0 Dec 10 19:34:35 crc kubenswrapper[4894]: I1210 19:34:35.453483 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerDied","Data":"575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62"} Dec 10 19:34:37 crc kubenswrapper[4894]: I1210 19:34:37.471170 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerStarted","Data":"aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f"} Dec 10 19:34:38 crc kubenswrapper[4894]: I1210 19:34:38.485721 4894 generic.go:334] "Generic (PLEG): container finished" podID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerID="aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f" exitCode=0 Dec 10 19:34:38 crc kubenswrapper[4894]: I1210 19:34:38.485785 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerDied","Data":"aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f"} Dec 10 19:34:39 crc kubenswrapper[4894]: I1210 19:34:39.506258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerStarted","Data":"614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c"} Dec 10 19:34:39 crc kubenswrapper[4894]: I1210 19:34:39.536459 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j2ch5" podStartSLOduration=2.805465263 podStartE2EDuration="6.536444025s" podCreationTimestamp="2025-12-10 19:34:33 +0000 UTC" firstStartedPulling="2025-12-10 19:34:35.456918942 +0000 UTC m=+2358.251766764" lastFinishedPulling="2025-12-10 19:34:39.187897714 +0000 UTC m=+2361.982745526" observedRunningTime="2025-12-10 19:34:39.527767719 +0000 UTC m=+2362.322615521" watchObservedRunningTime="2025-12-10 19:34:39.536444025 +0000 UTC m=+2362.331291817" Dec 10 19:34:43 crc kubenswrapper[4894]: I1210 19:34:43.541753 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:34:43 crc kubenswrapper[4894]: E1210 19:34:43.542307 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:34:43 crc kubenswrapper[4894]: I1210 19:34:43.674534 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:43 crc kubenswrapper[4894]: I1210 19:34:43.674592 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:43 crc kubenswrapper[4894]: I1210 19:34:43.737082 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:44 crc kubenswrapper[4894]: I1210 19:34:44.632812 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:44 crc kubenswrapper[4894]: I1210 19:34:44.695654 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j2ch5"] Dec 10 19:34:46 crc kubenswrapper[4894]: I1210 19:34:46.575830 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j2ch5" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="registry-server" containerID="cri-o://614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c" gracePeriod=2 Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.164952 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.274991 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-utilities\") pod \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.275153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnzz2\" (UniqueName: \"kubernetes.io/projected/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-kube-api-access-jnzz2\") pod \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.275403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-catalog-content\") pod \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\" (UID: \"f06c0268-818e-4ac2-a4b4-3785f1ff84f3\") " Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.276217 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-utilities" (OuterVolumeSpecName: "utilities") pod "f06c0268-818e-4ac2-a4b4-3785f1ff84f3" (UID: "f06c0268-818e-4ac2-a4b4-3785f1ff84f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.283125 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-kube-api-access-jnzz2" (OuterVolumeSpecName: "kube-api-access-jnzz2") pod "f06c0268-818e-4ac2-a4b4-3785f1ff84f3" (UID: "f06c0268-818e-4ac2-a4b4-3785f1ff84f3"). InnerVolumeSpecName "kube-api-access-jnzz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.341041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f06c0268-818e-4ac2-a4b4-3785f1ff84f3" (UID: "f06c0268-818e-4ac2-a4b4-3785f1ff84f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.378371 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.378404 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.378415 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnzz2\" (UniqueName: \"kubernetes.io/projected/f06c0268-818e-4ac2-a4b4-3785f1ff84f3-kube-api-access-jnzz2\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.607769 4894 generic.go:334] "Generic (PLEG): container finished" podID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerID="614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c" exitCode=0 Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.607822 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerDied","Data":"614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c"} Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.607912 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ch5" event={"ID":"f06c0268-818e-4ac2-a4b4-3785f1ff84f3","Type":"ContainerDied","Data":"a369e82e783fe82474cd5d993f24c6b61224f05ebfd82aafdd9ddecb22a90426"} Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.607939 4894 scope.go:117] "RemoveContainer" containerID="614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.608126 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ch5" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.641829 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j2ch5"] Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.650962 4894 scope.go:117] "RemoveContainer" containerID="aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.665063 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j2ch5"] Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.671196 4894 scope.go:117] "RemoveContainer" containerID="575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.716804 4894 scope.go:117] "RemoveContainer" containerID="614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c" Dec 10 19:34:47 crc kubenswrapper[4894]: E1210 19:34:47.717319 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c\": container with ID starting with 614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c not found: ID does not exist" containerID="614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.717415 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c"} err="failed to get container status \"614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c\": rpc error: code = NotFound desc = could not find container \"614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c\": container with ID starting with 614250ee669d2bf7e4e9d1cadf8edf21d33b6b4a68cf8e53f787cdb92380d98c not found: ID does not exist" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.717489 4894 scope.go:117] "RemoveContainer" containerID="aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f" Dec 10 19:34:47 crc kubenswrapper[4894]: E1210 19:34:47.717764 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f\": container with ID starting with aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f not found: ID does not exist" containerID="aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.717836 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f"} err="failed to get container status \"aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f\": rpc error: code = NotFound desc = could not find container \"aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f\": container with ID starting with aed5cc1392fda51ad039a0f8bd567dd961256b56c2b1aab46d37bc8ca456677f not found: ID does not exist" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.717937 4894 scope.go:117] "RemoveContainer" containerID="575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62" Dec 10 19:34:47 crc kubenswrapper[4894]: E1210 19:34:47.718379 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62\": container with ID starting with 575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62 not found: ID does not exist" containerID="575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62" Dec 10 19:34:47 crc kubenswrapper[4894]: I1210 19:34:47.718464 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62"} err="failed to get container status \"575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62\": rpc error: code = NotFound desc = could not find container \"575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62\": container with ID starting with 575b7cecb8e98179e20b8f85c8d802d11ab3d8f7a2aeaebc4b7811a12aadad62 not found: ID does not exist" Dec 10 19:34:49 crc kubenswrapper[4894]: I1210 19:34:49.558899 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" path="/var/lib/kubelet/pods/f06c0268-818e-4ac2-a4b4-3785f1ff84f3/volumes" Dec 10 19:34:55 crc kubenswrapper[4894]: I1210 19:34:55.540783 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:34:55 crc kubenswrapper[4894]: E1210 19:34:55.541808 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:35:10 crc kubenswrapper[4894]: I1210 19:35:10.541669 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:35:10 crc kubenswrapper[4894]: E1210 19:35:10.542931 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:35:24 crc kubenswrapper[4894]: I1210 19:35:24.541621 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:35:24 crc kubenswrapper[4894]: E1210 19:35:24.542663 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:35:35 crc kubenswrapper[4894]: I1210 19:35:35.543769 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:35:35 crc kubenswrapper[4894]: E1210 19:35:35.546635 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:35:50 crc kubenswrapper[4894]: I1210 19:35:50.541658 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:35:50 crc kubenswrapper[4894]: E1210 19:35:50.542542 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:36:03 crc kubenswrapper[4894]: I1210 19:36:03.541580 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:36:03 crc kubenswrapper[4894]: E1210 19:36:03.543872 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:36:17 crc kubenswrapper[4894]: I1210 19:36:17.567823 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:36:17 crc kubenswrapper[4894]: E1210 19:36:17.569256 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:36:29 crc kubenswrapper[4894]: I1210 19:36:29.541480 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:36:29 crc kubenswrapper[4894]: E1210 19:36:29.542980 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:36:41 crc kubenswrapper[4894]: I1210 19:36:41.541789 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:36:41 crc kubenswrapper[4894]: E1210 19:36:41.542698 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:36:55 crc kubenswrapper[4894]: I1210 19:36:55.541768 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:36:55 crc kubenswrapper[4894]: E1210 19:36:55.542785 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:37:06 crc kubenswrapper[4894]: I1210 19:37:06.541216 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:37:06 crc kubenswrapper[4894]: E1210 19:37:06.542036 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:37:21 crc kubenswrapper[4894]: I1210 19:37:21.542751 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:37:21 crc kubenswrapper[4894]: E1210 19:37:21.543598 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:37:33 crc kubenswrapper[4894]: I1210 19:37:33.543286 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:37:33 crc kubenswrapper[4894]: E1210 19:37:33.544018 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:37:46 crc kubenswrapper[4894]: I1210 19:37:46.541789 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:37:46 crc kubenswrapper[4894]: E1210 19:37:46.542642 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:38:00 crc kubenswrapper[4894]: I1210 19:38:00.541879 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:38:00 crc kubenswrapper[4894]: E1210 19:38:00.542693 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:38:14 crc kubenswrapper[4894]: I1210 19:38:14.541450 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:38:14 crc kubenswrapper[4894]: E1210 19:38:14.542343 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:38:25 crc kubenswrapper[4894]: I1210 19:38:25.541170 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:38:25 crc kubenswrapper[4894]: E1210 19:38:25.541987 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:38:31 crc kubenswrapper[4894]: I1210 19:38:31.097872 4894 generic.go:334] "Generic (PLEG): container finished" podID="a5f93c12-d3db-42db-b431-861c46535ba4" containerID="3ca07f4eb9bf413167d007f63917d8065fea7281e2c4596b7c1f6a16be6aeafe" exitCode=0 Dec 10 19:38:31 crc kubenswrapper[4894]: I1210 19:38:31.098010 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" event={"ID":"a5f93c12-d3db-42db-b431-861c46535ba4","Type":"ContainerDied","Data":"3ca07f4eb9bf413167d007f63917d8065fea7281e2c4596b7c1f6a16be6aeafe"} Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.581901 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.741373 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-combined-ca-bundle\") pod \"a5f93c12-d3db-42db-b431-861c46535ba4\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.741442 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-inventory\") pod \"a5f93c12-d3db-42db-b431-861c46535ba4\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.741565 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dfrn\" (UniqueName: \"kubernetes.io/projected/a5f93c12-d3db-42db-b431-861c46535ba4-kube-api-access-7dfrn\") pod \"a5f93c12-d3db-42db-b431-861c46535ba4\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.741661 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-secret-0\") pod \"a5f93c12-d3db-42db-b431-861c46535ba4\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.741710 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-ssh-key\") pod \"a5f93c12-d3db-42db-b431-861c46535ba4\" (UID: \"a5f93c12-d3db-42db-b431-861c46535ba4\") " Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.746467 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a5f93c12-d3db-42db-b431-861c46535ba4" (UID: "a5f93c12-d3db-42db-b431-861c46535ba4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.757065 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f93c12-d3db-42db-b431-861c46535ba4-kube-api-access-7dfrn" (OuterVolumeSpecName: "kube-api-access-7dfrn") pod "a5f93c12-d3db-42db-b431-861c46535ba4" (UID: "a5f93c12-d3db-42db-b431-861c46535ba4"). InnerVolumeSpecName "kube-api-access-7dfrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.768884 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a5f93c12-d3db-42db-b431-861c46535ba4" (UID: "a5f93c12-d3db-42db-b431-861c46535ba4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.777156 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-inventory" (OuterVolumeSpecName: "inventory") pod "a5f93c12-d3db-42db-b431-861c46535ba4" (UID: "a5f93c12-d3db-42db-b431-861c46535ba4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.779797 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "a5f93c12-d3db-42db-b431-861c46535ba4" (UID: "a5f93c12-d3db-42db-b431-861c46535ba4"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.844228 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.844269 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.844285 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.844297 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dfrn\" (UniqueName: \"kubernetes.io/projected/a5f93c12-d3db-42db-b431-861c46535ba4-kube-api-access-7dfrn\") on node \"crc\" DevicePath \"\"" Dec 10 19:38:32 crc kubenswrapper[4894]: I1210 19:38:32.844309 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a5f93c12-d3db-42db-b431-861c46535ba4-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.123438 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" event={"ID":"a5f93c12-d3db-42db-b431-861c46535ba4","Type":"ContainerDied","Data":"e5cb2c8d6de2f799e87f0a9b94e84ef9e0e5a26a0572578e14bf7c1dabf6c850"} Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.123474 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5cb2c8d6de2f799e87f0a9b94e84ef9e0e5a26a0572578e14bf7c1dabf6c850" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.123549 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.246130 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq"] Dec 10 19:38:33 crc kubenswrapper[4894]: E1210 19:38:33.246917 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="extract-content" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.246938 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="extract-content" Dec 10 19:38:33 crc kubenswrapper[4894]: E1210 19:38:33.246973 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="extract-utilities" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.246982 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="extract-utilities" Dec 10 19:38:33 crc kubenswrapper[4894]: E1210 19:38:33.247011 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f93c12-d3db-42db-b431-861c46535ba4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.247021 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f93c12-d3db-42db-b431-861c46535ba4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 19:38:33 crc kubenswrapper[4894]: E1210 19:38:33.247041 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="registry-server" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.247048 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="registry-server" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.247306 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f93c12-d3db-42db-b431-861c46535ba4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.247322 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f06c0268-818e-4ac2-a4b4-3785f1ff84f3" containerName="registry-server" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.248299 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.251011 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.251422 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.251738 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.252008 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.252307 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.252425 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.252311 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.264692 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq"] Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.355883 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvdj\" (UniqueName: \"kubernetes.io/projected/31e2c494-b1f4-453b-a536-26ac016c63f9-kube-api-access-qsvdj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356058 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356099 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356142 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356529 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356582 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.356638 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458440 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458489 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458517 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458575 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvdj\" (UniqueName: \"kubernetes.io/projected/31e2c494-b1f4-453b-a536-26ac016c63f9-kube-api-access-qsvdj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458632 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458655 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458686 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.458764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.459762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.463348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.463398 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.463599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.464288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.464488 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.465552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.469521 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.479359 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvdj\" (UniqueName: \"kubernetes.io/projected/31e2c494-b1f4-453b-a536-26ac016c63f9-kube-api-access-qsvdj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4dzqq\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:33 crc kubenswrapper[4894]: I1210 19:38:33.566129 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:38:34 crc kubenswrapper[4894]: I1210 19:38:34.110571 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq"] Dec 10 19:38:34 crc kubenswrapper[4894]: I1210 19:38:34.113211 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:38:34 crc kubenswrapper[4894]: I1210 19:38:34.134548 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" event={"ID":"31e2c494-b1f4-453b-a536-26ac016c63f9","Type":"ContainerStarted","Data":"a786c751fbed15705c0c98bac657515c3ef6af688acbeee67824ca7e07e6c848"} Dec 10 19:38:35 crc kubenswrapper[4894]: I1210 19:38:35.154033 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" event={"ID":"31e2c494-b1f4-453b-a536-26ac016c63f9","Type":"ContainerStarted","Data":"0991d061438d247489a510df882a8015339aec7bdcd1dd3466f38dee3e70d375"} Dec 10 19:38:35 crc kubenswrapper[4894]: I1210 19:38:35.234671 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" podStartSLOduration=1.787122599 podStartE2EDuration="2.234647761s" podCreationTimestamp="2025-12-10 19:38:33 +0000 UTC" firstStartedPulling="2025-12-10 19:38:34.113019346 +0000 UTC m=+2596.907867138" lastFinishedPulling="2025-12-10 19:38:34.560544508 +0000 UTC m=+2597.355392300" observedRunningTime="2025-12-10 19:38:35.203456411 +0000 UTC m=+2597.998304243" watchObservedRunningTime="2025-12-10 19:38:35.234647761 +0000 UTC m=+2598.029495563" Dec 10 19:38:40 crc kubenswrapper[4894]: I1210 19:38:40.541132 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:38:40 crc kubenswrapper[4894]: E1210 19:38:40.561788 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:38:52 crc kubenswrapper[4894]: I1210 19:38:52.541166 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:38:53 crc kubenswrapper[4894]: I1210 19:38:53.347364 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"1aa4892d7a0a41a8534e895a42a33a49b6475f64aedea14e0aa2010d0c36ab5e"} Dec 10 19:39:49 crc kubenswrapper[4894]: I1210 19:39:49.802326 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4bs5h"] Dec 10 19:39:49 crc kubenswrapper[4894]: I1210 19:39:49.805584 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:49 crc kubenswrapper[4894]: I1210 19:39:49.822605 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bs5h"] Dec 10 19:39:49 crc kubenswrapper[4894]: I1210 19:39:49.949526 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-utilities\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:49 crc kubenswrapper[4894]: I1210 19:39:49.950026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dncq5\" (UniqueName: \"kubernetes.io/projected/fc1affbc-d21e-40fc-b148-6566c4f61c48-kube-api-access-dncq5\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:49 crc kubenswrapper[4894]: I1210 19:39:49.950098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-catalog-content\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.051982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-catalog-content\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.052175 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-utilities\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.052229 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dncq5\" (UniqueName: \"kubernetes.io/projected/fc1affbc-d21e-40fc-b148-6566c4f61c48-kube-api-access-dncq5\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.052651 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-catalog-content\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.052766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-utilities\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.074871 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dncq5\" (UniqueName: \"kubernetes.io/projected/fc1affbc-d21e-40fc-b148-6566c4f61c48-kube-api-access-dncq5\") pod \"redhat-operators-4bs5h\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.128320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.616110 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bs5h"] Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.943673 4894 generic.go:334] "Generic (PLEG): container finished" podID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerID="1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e" exitCode=0 Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.943898 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerDied","Data":"1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e"} Dec 10 19:39:50 crc kubenswrapper[4894]: I1210 19:39:50.947941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerStarted","Data":"af2861112b2692115d66baf9ea0c4ed4c75889dbe6f25b0c64df20e865f04bb4"} Dec 10 19:39:52 crc kubenswrapper[4894]: I1210 19:39:52.963595 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerStarted","Data":"1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac"} Dec 10 19:39:56 crc kubenswrapper[4894]: I1210 19:39:56.000473 4894 generic.go:334] "Generic (PLEG): container finished" podID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerID="1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac" exitCode=0 Dec 10 19:39:56 crc kubenswrapper[4894]: I1210 19:39:56.000572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerDied","Data":"1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac"} Dec 10 19:39:58 crc kubenswrapper[4894]: I1210 19:39:58.036287 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerStarted","Data":"3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3"} Dec 10 19:39:58 crc kubenswrapper[4894]: I1210 19:39:58.063259 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4bs5h" podStartSLOduration=3.031213938 podStartE2EDuration="9.063241726s" podCreationTimestamp="2025-12-10 19:39:49 +0000 UTC" firstStartedPulling="2025-12-10 19:39:50.946213976 +0000 UTC m=+2673.741061768" lastFinishedPulling="2025-12-10 19:39:56.978241754 +0000 UTC m=+2679.773089556" observedRunningTime="2025-12-10 19:39:58.056317987 +0000 UTC m=+2680.851165799" watchObservedRunningTime="2025-12-10 19:39:58.063241726 +0000 UTC m=+2680.858089518" Dec 10 19:40:00 crc kubenswrapper[4894]: I1210 19:40:00.128608 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:40:00 crc kubenswrapper[4894]: I1210 19:40:00.129221 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:40:01 crc kubenswrapper[4894]: I1210 19:40:01.178346 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4bs5h" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="registry-server" probeResult="failure" output=< Dec 10 19:40:01 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:40:01 crc kubenswrapper[4894]: > Dec 10 19:40:10 crc kubenswrapper[4894]: I1210 19:40:10.193143 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:40:10 crc kubenswrapper[4894]: I1210 19:40:10.292488 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:40:10 crc kubenswrapper[4894]: I1210 19:40:10.442413 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bs5h"] Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.193451 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4bs5h" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="registry-server" containerID="cri-o://3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3" gracePeriod=2 Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.812502 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.939790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dncq5\" (UniqueName: \"kubernetes.io/projected/fc1affbc-d21e-40fc-b148-6566c4f61c48-kube-api-access-dncq5\") pod \"fc1affbc-d21e-40fc-b148-6566c4f61c48\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.939976 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-utilities\") pod \"fc1affbc-d21e-40fc-b148-6566c4f61c48\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.940128 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-catalog-content\") pod \"fc1affbc-d21e-40fc-b148-6566c4f61c48\" (UID: \"fc1affbc-d21e-40fc-b148-6566c4f61c48\") " Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.940943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-utilities" (OuterVolumeSpecName: "utilities") pod "fc1affbc-d21e-40fc-b148-6566c4f61c48" (UID: "fc1affbc-d21e-40fc-b148-6566c4f61c48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:40:12 crc kubenswrapper[4894]: I1210 19:40:12.945262 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1affbc-d21e-40fc-b148-6566c4f61c48-kube-api-access-dncq5" (OuterVolumeSpecName: "kube-api-access-dncq5") pod "fc1affbc-d21e-40fc-b148-6566c4f61c48" (UID: "fc1affbc-d21e-40fc-b148-6566c4f61c48"). InnerVolumeSpecName "kube-api-access-dncq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.042179 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dncq5\" (UniqueName: \"kubernetes.io/projected/fc1affbc-d21e-40fc-b148-6566c4f61c48-kube-api-access-dncq5\") on node \"crc\" DevicePath \"\"" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.042390 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.067383 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc1affbc-d21e-40fc-b148-6566c4f61c48" (UID: "fc1affbc-d21e-40fc-b148-6566c4f61c48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.146116 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1affbc-d21e-40fc-b148-6566c4f61c48-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.206698 4894 generic.go:334] "Generic (PLEG): container finished" podID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerID="3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3" exitCode=0 Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.206744 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerDied","Data":"3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3"} Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.206769 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bs5h" event={"ID":"fc1affbc-d21e-40fc-b148-6566c4f61c48","Type":"ContainerDied","Data":"af2861112b2692115d66baf9ea0c4ed4c75889dbe6f25b0c64df20e865f04bb4"} Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.206800 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bs5h" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.206818 4894 scope.go:117] "RemoveContainer" containerID="3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.237394 4894 scope.go:117] "RemoveContainer" containerID="1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.261539 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bs5h"] Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.278310 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4bs5h"] Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.296830 4894 scope.go:117] "RemoveContainer" containerID="1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.341405 4894 scope.go:117] "RemoveContainer" containerID="3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3" Dec 10 19:40:13 crc kubenswrapper[4894]: E1210 19:40:13.343364 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3\": container with ID starting with 3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3 not found: ID does not exist" containerID="3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.343407 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3"} err="failed to get container status \"3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3\": rpc error: code = NotFound desc = could not find container \"3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3\": container with ID starting with 3fae2db6cfe42787ec7c9cfb4b611a39d41d2f650b64c5ad40dc5b998385b1e3 not found: ID does not exist" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.343441 4894 scope.go:117] "RemoveContainer" containerID="1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac" Dec 10 19:40:13 crc kubenswrapper[4894]: E1210 19:40:13.343930 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac\": container with ID starting with 1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac not found: ID does not exist" containerID="1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.343977 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac"} err="failed to get container status \"1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac\": rpc error: code = NotFound desc = could not find container \"1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac\": container with ID starting with 1f80cb5f3b129ebc1d4aa17ad7c77d3937e9a748615bd4f7ec983d48576e5eac not found: ID does not exist" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.344011 4894 scope.go:117] "RemoveContainer" containerID="1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e" Dec 10 19:40:13 crc kubenswrapper[4894]: E1210 19:40:13.344398 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e\": container with ID starting with 1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e not found: ID does not exist" containerID="1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.344440 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e"} err="failed to get container status \"1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e\": rpc error: code = NotFound desc = could not find container \"1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e\": container with ID starting with 1beb3eab99c718de38e798cc1d9cc06d2df63d8a24dc3ead20604ee8bb9a739e not found: ID does not exist" Dec 10 19:40:13 crc kubenswrapper[4894]: I1210 19:40:13.558229 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" path="/var/lib/kubelet/pods/fc1affbc-d21e-40fc-b148-6566c4f61c48/volumes" Dec 10 19:41:19 crc kubenswrapper[4894]: I1210 19:41:19.577579 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:41:19 crc kubenswrapper[4894]: I1210 19:41:19.578241 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:41:37 crc kubenswrapper[4894]: I1210 19:41:37.080670 4894 generic.go:334] "Generic (PLEG): container finished" podID="31e2c494-b1f4-453b-a536-26ac016c63f9" containerID="0991d061438d247489a510df882a8015339aec7bdcd1dd3466f38dee3e70d375" exitCode=0 Dec 10 19:41:37 crc kubenswrapper[4894]: I1210 19:41:37.080811 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" event={"ID":"31e2c494-b1f4-453b-a536-26ac016c63f9","Type":"ContainerDied","Data":"0991d061438d247489a510df882a8015339aec7bdcd1dd3466f38dee3e70d375"} Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.695359 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.767861 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-1\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.767924 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvdj\" (UniqueName: \"kubernetes.io/projected/31e2c494-b1f4-453b-a536-26ac016c63f9-kube-api-access-qsvdj\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768035 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-combined-ca-bundle\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768337 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-ssh-key\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768501 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-0\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768556 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-extra-config-0\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768603 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-inventory\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768625 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-0\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.768659 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-1\") pod \"31e2c494-b1f4-453b-a536-26ac016c63f9\" (UID: \"31e2c494-b1f4-453b-a536-26ac016c63f9\") " Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.773726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.774266 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e2c494-b1f4-453b-a536-26ac016c63f9-kube-api-access-qsvdj" (OuterVolumeSpecName: "kube-api-access-qsvdj") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "kube-api-access-qsvdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.796357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.796356 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.797760 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.801209 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.808870 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-inventory" (OuterVolumeSpecName: "inventory") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.813765 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.817741 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31e2c494-b1f4-453b-a536-26ac016c63f9" (UID: "31e2c494-b1f4-453b-a536-26ac016c63f9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871028 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871079 4894 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871122 4894 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871136 4894 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871148 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvdj\" (UniqueName: \"kubernetes.io/projected/31e2c494-b1f4-453b-a536-26ac016c63f9-kube-api-access-qsvdj\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871159 4894 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871170 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871181 4894 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:38 crc kubenswrapper[4894]: I1210 19:41:38.871194 4894 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31e2c494-b1f4-453b-a536-26ac016c63f9-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.106350 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" event={"ID":"31e2c494-b1f4-453b-a536-26ac016c63f9","Type":"ContainerDied","Data":"a786c751fbed15705c0c98bac657515c3ef6af688acbeee67824ca7e07e6c848"} Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.106388 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4dzqq" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.106406 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a786c751fbed15705c0c98bac657515c3ef6af688acbeee67824ca7e07e6c848" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.249340 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59"] Dec 10 19:41:39 crc kubenswrapper[4894]: E1210 19:41:39.249810 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="extract-utilities" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.249830 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="extract-utilities" Dec 10 19:41:39 crc kubenswrapper[4894]: E1210 19:41:39.251034 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e2c494-b1f4-453b-a536-26ac016c63f9" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.251058 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e2c494-b1f4-453b-a536-26ac016c63f9" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 19:41:39 crc kubenswrapper[4894]: E1210 19:41:39.251101 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="extract-content" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.251112 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="extract-content" Dec 10 19:41:39 crc kubenswrapper[4894]: E1210 19:41:39.251126 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="registry-server" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.251134 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="registry-server" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.251456 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e2c494-b1f4-453b-a536-26ac016c63f9" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.251490 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1affbc-d21e-40fc-b148-6566c4f61c48" containerName="registry-server" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.252734 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.256242 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.257750 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.258079 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.258299 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.258553 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g7wt2" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.266808 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59"] Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.277839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjhwt\" (UniqueName: \"kubernetes.io/projected/08cd9bee-4229-441c-9eb5-4d6beea60c79-kube-api-access-cjhwt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.277949 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.278056 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.278083 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.278118 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.278171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.278196 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.378753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.378811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.378829 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.378924 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjhwt\" (UniqueName: \"kubernetes.io/projected/08cd9bee-4229-441c-9eb5-4d6beea60c79-kube-api-access-cjhwt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.378976 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.379037 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.379057 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.382817 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.383689 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.384003 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.384763 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.385331 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.396944 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.399820 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjhwt\" (UniqueName: \"kubernetes.io/projected/08cd9bee-4229-441c-9eb5-4d6beea60c79-kube-api-access-cjhwt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bsm59\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:39 crc kubenswrapper[4894]: I1210 19:41:39.586138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:41:40 crc kubenswrapper[4894]: I1210 19:41:40.218482 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59"] Dec 10 19:41:41 crc kubenswrapper[4894]: I1210 19:41:41.130834 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" event={"ID":"08cd9bee-4229-441c-9eb5-4d6beea60c79","Type":"ContainerStarted","Data":"508cea8b669f639629dec79bd14a35f1958bda0ac75abf2d69652e4b56663ad2"} Dec 10 19:41:41 crc kubenswrapper[4894]: I1210 19:41:41.131691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" event={"ID":"08cd9bee-4229-441c-9eb5-4d6beea60c79","Type":"ContainerStarted","Data":"70a3c3edec7861b6e8c7baebd6afa3b42b9cb5b046aadccd598824016a5dece5"} Dec 10 19:41:41 crc kubenswrapper[4894]: I1210 19:41:41.160312 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" podStartSLOduration=1.6609360149999999 podStartE2EDuration="2.160281002s" podCreationTimestamp="2025-12-10 19:41:39 +0000 UTC" firstStartedPulling="2025-12-10 19:41:40.221114056 +0000 UTC m=+2783.015961858" lastFinishedPulling="2025-12-10 19:41:40.720459042 +0000 UTC m=+2783.515306845" observedRunningTime="2025-12-10 19:41:41.156355035 +0000 UTC m=+2783.951202857" watchObservedRunningTime="2025-12-10 19:41:41.160281002 +0000 UTC m=+2783.955128834" Dec 10 19:41:49 crc kubenswrapper[4894]: I1210 19:41:49.576936 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:41:49 crc kubenswrapper[4894]: I1210 19:41:49.577584 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:42:19 crc kubenswrapper[4894]: I1210 19:42:19.577063 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:42:19 crc kubenswrapper[4894]: I1210 19:42:19.577927 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:42:19 crc kubenswrapper[4894]: I1210 19:42:19.578030 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:42:19 crc kubenswrapper[4894]: I1210 19:42:19.579055 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1aa4892d7a0a41a8534e895a42a33a49b6475f64aedea14e0aa2010d0c36ab5e"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:42:19 crc kubenswrapper[4894]: I1210 19:42:19.579118 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://1aa4892d7a0a41a8534e895a42a33a49b6475f64aedea14e0aa2010d0c36ab5e" gracePeriod=600 Dec 10 19:42:20 crc kubenswrapper[4894]: I1210 19:42:20.574540 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="1aa4892d7a0a41a8534e895a42a33a49b6475f64aedea14e0aa2010d0c36ab5e" exitCode=0 Dec 10 19:42:20 crc kubenswrapper[4894]: I1210 19:42:20.574661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"1aa4892d7a0a41a8534e895a42a33a49b6475f64aedea14e0aa2010d0c36ab5e"} Dec 10 19:42:20 crc kubenswrapper[4894]: I1210 19:42:20.575177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772"} Dec 10 19:42:20 crc kubenswrapper[4894]: I1210 19:42:20.575219 4894 scope.go:117] "RemoveContainer" containerID="b8280782845338885dd1a614e6aad811d698adbebf00768dc9d6098d66d1df8a" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.428930 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8z64s"] Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.432349 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.461125 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g94pw\" (UniqueName: \"kubernetes.io/projected/9272eef5-dbd0-4431-b9e0-e114459f1c86-kube-api-access-g94pw\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.461260 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-utilities\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.461414 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-catalog-content\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.464839 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z64s"] Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.564329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-catalog-content\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.564693 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-catalog-content\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.565337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g94pw\" (UniqueName: \"kubernetes.io/projected/9272eef5-dbd0-4431-b9e0-e114459f1c86-kube-api-access-g94pw\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.566490 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-utilities\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.566159 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-utilities\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.585915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g94pw\" (UniqueName: \"kubernetes.io/projected/9272eef5-dbd0-4431-b9e0-e114459f1c86-kube-api-access-g94pw\") pod \"certified-operators-8z64s\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:46 crc kubenswrapper[4894]: I1210 19:43:46.762684 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:47 crc kubenswrapper[4894]: I1210 19:43:47.267224 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z64s"] Dec 10 19:43:47 crc kubenswrapper[4894]: I1210 19:43:47.538581 4894 generic.go:334] "Generic (PLEG): container finished" podID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerID="e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3" exitCode=0 Dec 10 19:43:47 crc kubenswrapper[4894]: I1210 19:43:47.538698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerDied","Data":"e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3"} Dec 10 19:43:47 crc kubenswrapper[4894]: I1210 19:43:47.538828 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerStarted","Data":"b6cb88942db5faad09328f2dbf937cf93c0b623e21de533d4490f13eb2fbb56e"} Dec 10 19:43:47 crc kubenswrapper[4894]: I1210 19:43:47.540705 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:43:53 crc kubenswrapper[4894]: I1210 19:43:53.625434 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerStarted","Data":"9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b"} Dec 10 19:43:54 crc kubenswrapper[4894]: I1210 19:43:54.637274 4894 generic.go:334] "Generic (PLEG): container finished" podID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerID="9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b" exitCode=0 Dec 10 19:43:54 crc kubenswrapper[4894]: I1210 19:43:54.637394 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerDied","Data":"9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b"} Dec 10 19:43:56 crc kubenswrapper[4894]: I1210 19:43:56.663739 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerStarted","Data":"3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78"} Dec 10 19:43:56 crc kubenswrapper[4894]: I1210 19:43:56.687120 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8z64s" podStartSLOduration=2.631880857 podStartE2EDuration="10.687096274s" podCreationTimestamp="2025-12-10 19:43:46 +0000 UTC" firstStartedPulling="2025-12-10 19:43:47.540412777 +0000 UTC m=+2910.335260579" lastFinishedPulling="2025-12-10 19:43:55.595628164 +0000 UTC m=+2918.390475996" observedRunningTime="2025-12-10 19:43:56.683415914 +0000 UTC m=+2919.478263756" watchObservedRunningTime="2025-12-10 19:43:56.687096274 +0000 UTC m=+2919.481944086" Dec 10 19:43:56 crc kubenswrapper[4894]: I1210 19:43:56.763103 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:56 crc kubenswrapper[4894]: I1210 19:43:56.763269 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:43:57 crc kubenswrapper[4894]: I1210 19:43:57.818526 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8z64s" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="registry-server" probeResult="failure" output=< Dec 10 19:43:57 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:43:57 crc kubenswrapper[4894]: > Dec 10 19:44:06 crc kubenswrapper[4894]: I1210 19:44:06.815330 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:44:06 crc kubenswrapper[4894]: I1210 19:44:06.866785 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:44:07 crc kubenswrapper[4894]: I1210 19:44:07.049678 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z64s"] Dec 10 19:44:08 crc kubenswrapper[4894]: I1210 19:44:08.794359 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8z64s" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="registry-server" containerID="cri-o://3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78" gracePeriod=2 Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.325831 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.446529 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g94pw\" (UniqueName: \"kubernetes.io/projected/9272eef5-dbd0-4431-b9e0-e114459f1c86-kube-api-access-g94pw\") pod \"9272eef5-dbd0-4431-b9e0-e114459f1c86\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.446566 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-catalog-content\") pod \"9272eef5-dbd0-4431-b9e0-e114459f1c86\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.446603 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-utilities\") pod \"9272eef5-dbd0-4431-b9e0-e114459f1c86\" (UID: \"9272eef5-dbd0-4431-b9e0-e114459f1c86\") " Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.447943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-utilities" (OuterVolumeSpecName: "utilities") pod "9272eef5-dbd0-4431-b9e0-e114459f1c86" (UID: "9272eef5-dbd0-4431-b9e0-e114459f1c86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.453758 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9272eef5-dbd0-4431-b9e0-e114459f1c86-kube-api-access-g94pw" (OuterVolumeSpecName: "kube-api-access-g94pw") pod "9272eef5-dbd0-4431-b9e0-e114459f1c86" (UID: "9272eef5-dbd0-4431-b9e0-e114459f1c86"). InnerVolumeSpecName "kube-api-access-g94pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.497556 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9272eef5-dbd0-4431-b9e0-e114459f1c86" (UID: "9272eef5-dbd0-4431-b9e0-e114459f1c86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.548596 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g94pw\" (UniqueName: \"kubernetes.io/projected/9272eef5-dbd0-4431-b9e0-e114459f1c86-kube-api-access-g94pw\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.548626 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.548638 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9272eef5-dbd0-4431-b9e0-e114459f1c86-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.807498 4894 generic.go:334] "Generic (PLEG): container finished" podID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerID="3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78" exitCode=0 Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.807558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerDied","Data":"3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78"} Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.807605 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z64s" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.807671 4894 scope.go:117] "RemoveContainer" containerID="3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.807622 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z64s" event={"ID":"9272eef5-dbd0-4431-b9e0-e114459f1c86","Type":"ContainerDied","Data":"b6cb88942db5faad09328f2dbf937cf93c0b623e21de533d4490f13eb2fbb56e"} Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.840089 4894 scope.go:117] "RemoveContainer" containerID="9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.846407 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z64s"] Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.865668 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8z64s"] Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.890591 4894 scope.go:117] "RemoveContainer" containerID="e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.933387 4894 scope.go:117] "RemoveContainer" containerID="3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78" Dec 10 19:44:09 crc kubenswrapper[4894]: E1210 19:44:09.933819 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78\": container with ID starting with 3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78 not found: ID does not exist" containerID="3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.933869 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78"} err="failed to get container status \"3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78\": rpc error: code = NotFound desc = could not find container \"3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78\": container with ID starting with 3e5fadf4a6c9e0be1a053e1839fa21d2ea1d61314e1c2f0bf971cc780c10be78 not found: ID does not exist" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.933890 4894 scope.go:117] "RemoveContainer" containerID="9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b" Dec 10 19:44:09 crc kubenswrapper[4894]: E1210 19:44:09.934213 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b\": container with ID starting with 9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b not found: ID does not exist" containerID="9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.934236 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b"} err="failed to get container status \"9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b\": rpc error: code = NotFound desc = could not find container \"9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b\": container with ID starting with 9e4fb942c23c5c4d4826eaf8d88de41a6f5e771e38e5a13089859fd1f8b1349b not found: ID does not exist" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.934259 4894 scope.go:117] "RemoveContainer" containerID="e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3" Dec 10 19:44:09 crc kubenswrapper[4894]: E1210 19:44:09.934607 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3\": container with ID starting with e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3 not found: ID does not exist" containerID="e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3" Dec 10 19:44:09 crc kubenswrapper[4894]: I1210 19:44:09.934674 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3"} err="failed to get container status \"e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3\": rpc error: code = NotFound desc = could not find container \"e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3\": container with ID starting with e443101c9209b222d0197ea2d79c17f8ebd2fd885faf0b5d2bbcaf5184d6a5e3 not found: ID does not exist" Dec 10 19:44:11 crc kubenswrapper[4894]: I1210 19:44:11.554245 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" path="/var/lib/kubelet/pods/9272eef5-dbd0-4431-b9e0-e114459f1c86/volumes" Dec 10 19:44:12 crc kubenswrapper[4894]: I1210 19:44:12.849142 4894 generic.go:334] "Generic (PLEG): container finished" podID="08cd9bee-4229-441c-9eb5-4d6beea60c79" containerID="508cea8b669f639629dec79bd14a35f1958bda0ac75abf2d69652e4b56663ad2" exitCode=0 Dec 10 19:44:12 crc kubenswrapper[4894]: I1210 19:44:12.849220 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" event={"ID":"08cd9bee-4229-441c-9eb5-4d6beea60c79","Type":"ContainerDied","Data":"508cea8b669f639629dec79bd14a35f1958bda0ac75abf2d69652e4b56663ad2"} Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.447694 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645041 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-2\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645150 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-1\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645190 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ssh-key\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645237 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjhwt\" (UniqueName: \"kubernetes.io/projected/08cd9bee-4229-441c-9eb5-4d6beea60c79-kube-api-access-cjhwt\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645282 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-telemetry-combined-ca-bundle\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645389 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-0\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.645413 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-inventory\") pod \"08cd9bee-4229-441c-9eb5-4d6beea60c79\" (UID: \"08cd9bee-4229-441c-9eb5-4d6beea60c79\") " Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.651790 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.664714 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08cd9bee-4229-441c-9eb5-4d6beea60c79-kube-api-access-cjhwt" (OuterVolumeSpecName: "kube-api-access-cjhwt") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "kube-api-access-cjhwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.674766 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.678348 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-inventory" (OuterVolumeSpecName: "inventory") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.679530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.680265 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.680755 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08cd9bee-4229-441c-9eb5-4d6beea60c79" (UID: "08cd9bee-4229-441c-9eb5-4d6beea60c79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747883 4894 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747915 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747926 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747937 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747953 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747965 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cd9bee-4229-441c-9eb5-4d6beea60c79-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.747976 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjhwt\" (UniqueName: \"kubernetes.io/projected/08cd9bee-4229-441c-9eb5-4d6beea60c79-kube-api-access-cjhwt\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.875946 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" event={"ID":"08cd9bee-4229-441c-9eb5-4d6beea60c79","Type":"ContainerDied","Data":"70a3c3edec7861b6e8c7baebd6afa3b42b9cb5b046aadccd598824016a5dece5"} Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.876005 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70a3c3edec7861b6e8c7baebd6afa3b42b9cb5b046aadccd598824016a5dece5" Dec 10 19:44:14 crc kubenswrapper[4894]: I1210 19:44:14.876043 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bsm59" Dec 10 19:44:19 crc kubenswrapper[4894]: I1210 19:44:19.577496 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:44:19 crc kubenswrapper[4894]: I1210 19:44:19.578077 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.258025 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j6dwk"] Dec 10 19:44:37 crc kubenswrapper[4894]: E1210 19:44:37.258970 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="extract-content" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.258983 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="extract-content" Dec 10 19:44:37 crc kubenswrapper[4894]: E1210 19:44:37.259011 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="extract-utilities" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.259017 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="extract-utilities" Dec 10 19:44:37 crc kubenswrapper[4894]: E1210 19:44:37.259033 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08cd9bee-4229-441c-9eb5-4d6beea60c79" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.259040 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="08cd9bee-4229-441c-9eb5-4d6beea60c79" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 19:44:37 crc kubenswrapper[4894]: E1210 19:44:37.259055 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="registry-server" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.259060 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="registry-server" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.259255 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="08cd9bee-4229-441c-9eb5-4d6beea60c79" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.259279 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9272eef5-dbd0-4431-b9e0-e114459f1c86" containerName="registry-server" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.260742 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.273816 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j6dwk"] Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.356396 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-catalog-content\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.356617 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zd9p\" (UniqueName: \"kubernetes.io/projected/b5d5dc5e-5fbb-4613-9470-622070ba482e-kube-api-access-4zd9p\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.356776 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-utilities\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.479885 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zd9p\" (UniqueName: \"kubernetes.io/projected/b5d5dc5e-5fbb-4613-9470-622070ba482e-kube-api-access-4zd9p\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.480035 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-utilities\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.480073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-catalog-content\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.480552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-catalog-content\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.480865 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-utilities\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.526886 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zd9p\" (UniqueName: \"kubernetes.io/projected/b5d5dc5e-5fbb-4613-9470-622070ba482e-kube-api-access-4zd9p\") pod \"community-operators-j6dwk\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:37 crc kubenswrapper[4894]: I1210 19:44:37.587451 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:38 crc kubenswrapper[4894]: I1210 19:44:38.098686 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j6dwk"] Dec 10 19:44:38 crc kubenswrapper[4894]: I1210 19:44:38.141410 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerStarted","Data":"3f3f9afe720bf640591e4709a566ce02714eef183fbb437f6a12459ba4704c0d"} Dec 10 19:44:39 crc kubenswrapper[4894]: I1210 19:44:39.151156 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerID="0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d" exitCode=0 Dec 10 19:44:39 crc kubenswrapper[4894]: I1210 19:44:39.151253 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerDied","Data":"0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d"} Dec 10 19:44:40 crc kubenswrapper[4894]: I1210 19:44:40.162668 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerStarted","Data":"9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1"} Dec 10 19:44:41 crc kubenswrapper[4894]: I1210 19:44:41.174836 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerID="9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1" exitCode=0 Dec 10 19:44:41 crc kubenswrapper[4894]: I1210 19:44:41.174947 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerDied","Data":"9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1"} Dec 10 19:44:42 crc kubenswrapper[4894]: I1210 19:44:42.201767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerStarted","Data":"c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1"} Dec 10 19:44:42 crc kubenswrapper[4894]: I1210 19:44:42.230303 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j6dwk" podStartSLOduration=2.499567689 podStartE2EDuration="5.230277012s" podCreationTimestamp="2025-12-10 19:44:37 +0000 UTC" firstStartedPulling="2025-12-10 19:44:39.154212171 +0000 UTC m=+2961.949059963" lastFinishedPulling="2025-12-10 19:44:41.884921484 +0000 UTC m=+2964.679769286" observedRunningTime="2025-12-10 19:44:42.226918762 +0000 UTC m=+2965.021766584" watchObservedRunningTime="2025-12-10 19:44:42.230277012 +0000 UTC m=+2965.025124804" Dec 10 19:44:47 crc kubenswrapper[4894]: I1210 19:44:47.588251 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:47 crc kubenswrapper[4894]: I1210 19:44:47.589000 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:47 crc kubenswrapper[4894]: I1210 19:44:47.652769 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:48 crc kubenswrapper[4894]: I1210 19:44:48.369494 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:48 crc kubenswrapper[4894]: I1210 19:44:48.453399 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j6dwk"] Dec 10 19:44:49 crc kubenswrapper[4894]: I1210 19:44:49.576783 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:44:49 crc kubenswrapper[4894]: I1210 19:44:49.576840 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.307590 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j6dwk" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="registry-server" containerID="cri-o://c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1" gracePeriod=2 Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.861037 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.891896 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-catalog-content\") pod \"b5d5dc5e-5fbb-4613-9470-622070ba482e\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.892088 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zd9p\" (UniqueName: \"kubernetes.io/projected/b5d5dc5e-5fbb-4613-9470-622070ba482e-kube-api-access-4zd9p\") pod \"b5d5dc5e-5fbb-4613-9470-622070ba482e\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.901074 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d5dc5e-5fbb-4613-9470-622070ba482e-kube-api-access-4zd9p" (OuterVolumeSpecName: "kube-api-access-4zd9p") pod "b5d5dc5e-5fbb-4613-9470-622070ba482e" (UID: "b5d5dc5e-5fbb-4613-9470-622070ba482e"). InnerVolumeSpecName "kube-api-access-4zd9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.947415 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5d5dc5e-5fbb-4613-9470-622070ba482e" (UID: "b5d5dc5e-5fbb-4613-9470-622070ba482e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.993876 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-utilities\") pod \"b5d5dc5e-5fbb-4613-9470-622070ba482e\" (UID: \"b5d5dc5e-5fbb-4613-9470-622070ba482e\") " Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.994210 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.994228 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zd9p\" (UniqueName: \"kubernetes.io/projected/b5d5dc5e-5fbb-4613-9470-622070ba482e-kube-api-access-4zd9p\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:50 crc kubenswrapper[4894]: I1210 19:44:50.995013 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-utilities" (OuterVolumeSpecName: "utilities") pod "b5d5dc5e-5fbb-4613-9470-622070ba482e" (UID: "b5d5dc5e-5fbb-4613-9470-622070ba482e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.095628 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5d5dc5e-5fbb-4613-9470-622070ba482e-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.319438 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerID="c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1" exitCode=0 Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.319486 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerDied","Data":"c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1"} Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.319514 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6dwk" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.319521 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6dwk" event={"ID":"b5d5dc5e-5fbb-4613-9470-622070ba482e","Type":"ContainerDied","Data":"3f3f9afe720bf640591e4709a566ce02714eef183fbb437f6a12459ba4704c0d"} Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.319577 4894 scope.go:117] "RemoveContainer" containerID="c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.356900 4894 scope.go:117] "RemoveContainer" containerID="9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.362401 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j6dwk"] Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.375408 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j6dwk"] Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.389320 4894 scope.go:117] "RemoveContainer" containerID="0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.445452 4894 scope.go:117] "RemoveContainer" containerID="c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1" Dec 10 19:44:51 crc kubenswrapper[4894]: E1210 19:44:51.445940 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1\": container with ID starting with c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1 not found: ID does not exist" containerID="c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.445983 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1"} err="failed to get container status \"c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1\": rpc error: code = NotFound desc = could not find container \"c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1\": container with ID starting with c787006194f63b515f977003be7f96e6fc3ca730e58fe1325222463b528db4e1 not found: ID does not exist" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.446013 4894 scope.go:117] "RemoveContainer" containerID="9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1" Dec 10 19:44:51 crc kubenswrapper[4894]: E1210 19:44:51.446558 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1\": container with ID starting with 9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1 not found: ID does not exist" containerID="9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.446589 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1"} err="failed to get container status \"9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1\": rpc error: code = NotFound desc = could not find container \"9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1\": container with ID starting with 9b94af4853847db1299e293e813c1d270db9904bab854bc50860c408c4259af1 not found: ID does not exist" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.446611 4894 scope.go:117] "RemoveContainer" containerID="0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d" Dec 10 19:44:51 crc kubenswrapper[4894]: E1210 19:44:51.447424 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d\": container with ID starting with 0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d not found: ID does not exist" containerID="0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.447457 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d"} err="failed to get container status \"0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d\": rpc error: code = NotFound desc = could not find container \"0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d\": container with ID starting with 0446b2eadb35d86ec137a71469b6ba6c32653005b7bf46f7566183dc80d3697d not found: ID does not exist" Dec 10 19:44:51 crc kubenswrapper[4894]: I1210 19:44:51.556237 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" path="/var/lib/kubelet/pods/b5d5dc5e-5fbb-4613-9470-622070ba482e/volumes" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.150578 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b"] Dec 10 19:45:00 crc kubenswrapper[4894]: E1210 19:45:00.152012 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="registry-server" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.152036 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="registry-server" Dec 10 19:45:00 crc kubenswrapper[4894]: E1210 19:45:00.152054 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="extract-utilities" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.152066 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="extract-utilities" Dec 10 19:45:00 crc kubenswrapper[4894]: E1210 19:45:00.152085 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="extract-content" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.152096 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="extract-content" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.152515 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d5dc5e-5fbb-4613-9470-622070ba482e" containerName="registry-server" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.153730 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.156475 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.156822 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.161068 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b"] Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.275976 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-config-volume\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.276080 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-secret-volume\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.276378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnz8m\" (UniqueName: \"kubernetes.io/projected/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-kube-api-access-cnz8m\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.378484 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-config-volume\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.379022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-secret-volume\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.379189 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnz8m\" (UniqueName: \"kubernetes.io/projected/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-kube-api-access-cnz8m\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.380046 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-config-volume\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.385049 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-secret-volume\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.406120 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnz8m\" (UniqueName: \"kubernetes.io/projected/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-kube-api-access-cnz8m\") pod \"collect-profiles-29423265-w8m6b\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.475983 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:00 crc kubenswrapper[4894]: I1210 19:45:00.971451 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b"] Dec 10 19:45:01 crc kubenswrapper[4894]: I1210 19:45:01.434123 4894 generic.go:334] "Generic (PLEG): container finished" podID="233e7ba3-2c3d-4f32-9e87-ecd59cadc624" containerID="1ea5e60cd65348c9823667378d3f82ceacbd7e2b77ba901b36f561b474e5ae32" exitCode=0 Dec 10 19:45:01 crc kubenswrapper[4894]: I1210 19:45:01.434177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" event={"ID":"233e7ba3-2c3d-4f32-9e87-ecd59cadc624","Type":"ContainerDied","Data":"1ea5e60cd65348c9823667378d3f82ceacbd7e2b77ba901b36f561b474e5ae32"} Dec 10 19:45:01 crc kubenswrapper[4894]: I1210 19:45:01.434438 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" event={"ID":"233e7ba3-2c3d-4f32-9e87-ecd59cadc624","Type":"ContainerStarted","Data":"de17eaa46aec4882d4ef9abc79afd142b61753c6c940caa13d0f5a190c1fe3f7"} Dec 10 19:45:02 crc kubenswrapper[4894]: I1210 19:45:02.939762 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.044467 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnz8m\" (UniqueName: \"kubernetes.io/projected/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-kube-api-access-cnz8m\") pod \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.044574 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-secret-volume\") pod \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.044658 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-config-volume\") pod \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\" (UID: \"233e7ba3-2c3d-4f32-9e87-ecd59cadc624\") " Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.046642 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-config-volume" (OuterVolumeSpecName: "config-volume") pod "233e7ba3-2c3d-4f32-9e87-ecd59cadc624" (UID: "233e7ba3-2c3d-4f32-9e87-ecd59cadc624"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.051720 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "233e7ba3-2c3d-4f32-9e87-ecd59cadc624" (UID: "233e7ba3-2c3d-4f32-9e87-ecd59cadc624"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.053534 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-kube-api-access-cnz8m" (OuterVolumeSpecName: "kube-api-access-cnz8m") pod "233e7ba3-2c3d-4f32-9e87-ecd59cadc624" (UID: "233e7ba3-2c3d-4f32-9e87-ecd59cadc624"). InnerVolumeSpecName "kube-api-access-cnz8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.146767 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnz8m\" (UniqueName: \"kubernetes.io/projected/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-kube-api-access-cnz8m\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.147140 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.147155 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/233e7ba3-2c3d-4f32-9e87-ecd59cadc624-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.467271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" event={"ID":"233e7ba3-2c3d-4f32-9e87-ecd59cadc624","Type":"ContainerDied","Data":"de17eaa46aec4882d4ef9abc79afd142b61753c6c940caa13d0f5a190c1fe3f7"} Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.467315 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de17eaa46aec4882d4ef9abc79afd142b61753c6c940caa13d0f5a190c1fe3f7" Dec 10 19:45:03 crc kubenswrapper[4894]: I1210 19:45:03.467331 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-w8m6b" Dec 10 19:45:04 crc kubenswrapper[4894]: I1210 19:45:04.014891 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp"] Dec 10 19:45:04 crc kubenswrapper[4894]: I1210 19:45:04.024203 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-2zvnp"] Dec 10 19:45:05 crc kubenswrapper[4894]: I1210 19:45:05.653879 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e789ccb-359b-4c16-a8f5-25864f34778c" path="/var/lib/kubelet/pods/6e789ccb-359b-4c16-a8f5-25864f34778c/volumes" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.231483 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 19:45:13 crc kubenswrapper[4894]: E1210 19:45:13.232692 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233e7ba3-2c3d-4f32-9e87-ecd59cadc624" containerName="collect-profiles" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.232708 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="233e7ba3-2c3d-4f32-9e87-ecd59cadc624" containerName="collect-profiles" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.233032 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="233e7ba3-2c3d-4f32-9e87-ecd59cadc624" containerName="collect-profiles" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.233961 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.242839 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.243172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.243397 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.243747 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-c8vz7" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.248026 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.338548 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339033 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339060 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhj5w\" (UniqueName: \"kubernetes.io/projected/1239b753-aad1-41b8-8df4-62df779823e5-kube-api-access-rhj5w\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339137 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339255 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339484 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339558 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.339673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-config-data\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441257 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441308 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhj5w\" (UniqueName: \"kubernetes.io/projected/1239b753-aad1-41b8-8df4-62df779823e5-kube-api-access-rhj5w\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441343 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441445 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-config-data\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.441505 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.442584 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.442622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.442731 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.443520 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.443671 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-config-data\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.448062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.454948 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.456677 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.461522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhj5w\" (UniqueName: \"kubernetes.io/projected/1239b753-aad1-41b8-8df4-62df779823e5-kube-api-access-rhj5w\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.489803 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " pod="openstack/tempest-tests-tempest" Dec 10 19:45:13 crc kubenswrapper[4894]: I1210 19:45:13.607725 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 19:45:14 crc kubenswrapper[4894]: I1210 19:45:14.107157 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 19:45:14 crc kubenswrapper[4894]: I1210 19:45:14.567820 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1239b753-aad1-41b8-8df4-62df779823e5","Type":"ContainerStarted","Data":"05855e15cfccb5fb0a438ccc888a3fcbbf260272de05f5e8a4ceec94e9bc392c"} Dec 10 19:45:19 crc kubenswrapper[4894]: I1210 19:45:19.577441 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:45:19 crc kubenswrapper[4894]: I1210 19:45:19.577992 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:45:19 crc kubenswrapper[4894]: I1210 19:45:19.578043 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:45:19 crc kubenswrapper[4894]: I1210 19:45:19.578620 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:45:19 crc kubenswrapper[4894]: I1210 19:45:19.578714 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" gracePeriod=600 Dec 10 19:45:19 crc kubenswrapper[4894]: E1210 19:45:19.699151 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:45:20 crc kubenswrapper[4894]: I1210 19:45:20.632366 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" exitCode=0 Dec 10 19:45:20 crc kubenswrapper[4894]: I1210 19:45:20.632510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772"} Dec 10 19:45:20 crc kubenswrapper[4894]: I1210 19:45:20.632766 4894 scope.go:117] "RemoveContainer" containerID="1aa4892d7a0a41a8534e895a42a33a49b6475f64aedea14e0aa2010d0c36ab5e" Dec 10 19:45:20 crc kubenswrapper[4894]: I1210 19:45:20.633838 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:45:20 crc kubenswrapper[4894]: E1210 19:45:20.634517 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:45:23 crc kubenswrapper[4894]: I1210 19:45:23.379277 4894 scope.go:117] "RemoveContainer" containerID="efd9770e718f824bf8d557c632532fb5fadb807e233864fca99663a3563c4df9" Dec 10 19:45:35 crc kubenswrapper[4894]: I1210 19:45:35.544294 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:45:35 crc kubenswrapper[4894]: E1210 19:45:35.545123 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:45:50 crc kubenswrapper[4894]: I1210 19:45:50.542468 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:45:50 crc kubenswrapper[4894]: E1210 19:45:50.543726 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:45:52 crc kubenswrapper[4894]: E1210 19:45:52.476982 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 10 19:45:52 crc kubenswrapper[4894]: E1210 19:45:52.477744 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rhj5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1239b753-aad1-41b8-8df4-62df779823e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:45:52 crc kubenswrapper[4894]: E1210 19:45:52.479454 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1239b753-aad1-41b8-8df4-62df779823e5" Dec 10 19:45:53 crc kubenswrapper[4894]: E1210 19:45:53.022711 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1239b753-aad1-41b8-8df4-62df779823e5" Dec 10 19:46:04 crc kubenswrapper[4894]: I1210 19:46:04.080299 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 19:46:04 crc kubenswrapper[4894]: I1210 19:46:04.542261 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:46:04 crc kubenswrapper[4894]: E1210 19:46:04.542880 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:46:06 crc kubenswrapper[4894]: I1210 19:46:06.155802 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1239b753-aad1-41b8-8df4-62df779823e5","Type":"ContainerStarted","Data":"9ae27fdf3ada8622de58120a31915c881237dc5b11abb61fd6d3e2d7792d95f8"} Dec 10 19:46:06 crc kubenswrapper[4894]: I1210 19:46:06.178000 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.208608919 podStartE2EDuration="54.177982161s" podCreationTimestamp="2025-12-10 19:45:12 +0000 UTC" firstStartedPulling="2025-12-10 19:45:14.107835995 +0000 UTC m=+2996.902683787" lastFinishedPulling="2025-12-10 19:46:04.077209237 +0000 UTC m=+3046.872057029" observedRunningTime="2025-12-10 19:46:06.171189128 +0000 UTC m=+3048.966036930" watchObservedRunningTime="2025-12-10 19:46:06.177982161 +0000 UTC m=+3048.972829953" Dec 10 19:46:17 crc kubenswrapper[4894]: I1210 19:46:17.549571 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:46:17 crc kubenswrapper[4894]: E1210 19:46:17.551229 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:46:30 crc kubenswrapper[4894]: I1210 19:46:30.541911 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:46:30 crc kubenswrapper[4894]: E1210 19:46:30.543265 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:46:41 crc kubenswrapper[4894]: I1210 19:46:41.541566 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:46:41 crc kubenswrapper[4894]: E1210 19:46:41.542240 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:46:52 crc kubenswrapper[4894]: I1210 19:46:52.542382 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:46:52 crc kubenswrapper[4894]: E1210 19:46:52.543181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:47:06 crc kubenswrapper[4894]: I1210 19:47:06.541906 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:47:06 crc kubenswrapper[4894]: E1210 19:47:06.543181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:47:20 crc kubenswrapper[4894]: I1210 19:47:20.544686 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:47:20 crc kubenswrapper[4894]: E1210 19:47:20.545811 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:47:33 crc kubenswrapper[4894]: I1210 19:47:33.541585 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:47:33 crc kubenswrapper[4894]: E1210 19:47:33.542521 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:47:46 crc kubenswrapper[4894]: I1210 19:47:46.541711 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:47:46 crc kubenswrapper[4894]: E1210 19:47:46.542575 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:47:57 crc kubenswrapper[4894]: I1210 19:47:57.548323 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:47:57 crc kubenswrapper[4894]: E1210 19:47:57.549273 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:48:12 crc kubenswrapper[4894]: I1210 19:48:12.541130 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:48:12 crc kubenswrapper[4894]: E1210 19:48:12.541962 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:48:27 crc kubenswrapper[4894]: I1210 19:48:27.560228 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:48:27 crc kubenswrapper[4894]: E1210 19:48:27.561108 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:48:40 crc kubenswrapper[4894]: I1210 19:48:40.540833 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:48:40 crc kubenswrapper[4894]: E1210 19:48:40.542464 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:48:54 crc kubenswrapper[4894]: I1210 19:48:54.541656 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:48:54 crc kubenswrapper[4894]: E1210 19:48:54.544084 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:49:08 crc kubenswrapper[4894]: I1210 19:49:08.543032 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:49:08 crc kubenswrapper[4894]: E1210 19:49:08.544535 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:49:22 crc kubenswrapper[4894]: I1210 19:49:22.542197 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:49:22 crc kubenswrapper[4894]: E1210 19:49:22.543077 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:49:36 crc kubenswrapper[4894]: I1210 19:49:36.541672 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:49:36 crc kubenswrapper[4894]: E1210 19:49:36.542380 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:49:48 crc kubenswrapper[4894]: I1210 19:49:48.541501 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:49:48 crc kubenswrapper[4894]: E1210 19:49:48.542695 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:49:49 crc kubenswrapper[4894]: I1210 19:49:49.788367 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qs977"] Dec 10 19:49:49 crc kubenswrapper[4894]: I1210 19:49:49.791022 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:49 crc kubenswrapper[4894]: I1210 19:49:49.798694 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs977"] Dec 10 19:49:49 crc kubenswrapper[4894]: I1210 19:49:49.984024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-utilities\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:49 crc kubenswrapper[4894]: I1210 19:49:49.984227 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8jsh\" (UniqueName: \"kubernetes.io/projected/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-kube-api-access-n8jsh\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:49 crc kubenswrapper[4894]: I1210 19:49:49.984368 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-catalog-content\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.086762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-utilities\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.087169 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8jsh\" (UniqueName: \"kubernetes.io/projected/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-kube-api-access-n8jsh\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.087326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-catalog-content\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.087388 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-utilities\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.087882 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-catalog-content\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.109156 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8jsh\" (UniqueName: \"kubernetes.io/projected/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-kube-api-access-n8jsh\") pod \"redhat-marketplace-qs977\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.125409 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:49:50 crc kubenswrapper[4894]: I1210 19:49:50.653372 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs977"] Dec 10 19:49:51 crc kubenswrapper[4894]: I1210 19:49:51.429274 4894 generic.go:334] "Generic (PLEG): container finished" podID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerID="19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12" exitCode=0 Dec 10 19:49:51 crc kubenswrapper[4894]: I1210 19:49:51.429416 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerDied","Data":"19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12"} Dec 10 19:49:51 crc kubenswrapper[4894]: I1210 19:49:51.429656 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerStarted","Data":"2b2df8025be0c35ba0f53737c85ef5642c67915e37889090416fa6cceee1ff9c"} Dec 10 19:49:51 crc kubenswrapper[4894]: I1210 19:49:51.436601 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:49:52 crc kubenswrapper[4894]: I1210 19:49:52.442048 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerStarted","Data":"2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55"} Dec 10 19:49:53 crc kubenswrapper[4894]: I1210 19:49:53.453076 4894 generic.go:334] "Generic (PLEG): container finished" podID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerID="2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55" exitCode=0 Dec 10 19:49:53 crc kubenswrapper[4894]: I1210 19:49:53.453167 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerDied","Data":"2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55"} Dec 10 19:49:54 crc kubenswrapper[4894]: I1210 19:49:54.467168 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerStarted","Data":"69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e"} Dec 10 19:49:54 crc kubenswrapper[4894]: I1210 19:49:54.493046 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qs977" podStartSLOduration=3.009387224 podStartE2EDuration="5.493021477s" podCreationTimestamp="2025-12-10 19:49:49 +0000 UTC" firstStartedPulling="2025-12-10 19:49:51.436192425 +0000 UTC m=+3274.231040217" lastFinishedPulling="2025-12-10 19:49:53.919826668 +0000 UTC m=+3276.714674470" observedRunningTime="2025-12-10 19:49:54.490040357 +0000 UTC m=+3277.284888179" watchObservedRunningTime="2025-12-10 19:49:54.493021477 +0000 UTC m=+3277.287869259" Dec 10 19:50:00 crc kubenswrapper[4894]: I1210 19:50:00.126574 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:50:00 crc kubenswrapper[4894]: I1210 19:50:00.128646 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:50:00 crc kubenswrapper[4894]: I1210 19:50:00.176534 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:50:00 crc kubenswrapper[4894]: I1210 19:50:00.573134 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:50:00 crc kubenswrapper[4894]: I1210 19:50:00.627284 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs977"] Dec 10 19:50:02 crc kubenswrapper[4894]: I1210 19:50:02.541650 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:50:02 crc kubenswrapper[4894]: E1210 19:50:02.542514 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:50:02 crc kubenswrapper[4894]: I1210 19:50:02.550443 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qs977" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="registry-server" containerID="cri-o://69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e" gracePeriod=2 Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.200126 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.296327 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-catalog-content\") pod \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.296488 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8jsh\" (UniqueName: \"kubernetes.io/projected/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-kube-api-access-n8jsh\") pod \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.297384 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-utilities\") pod \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\" (UID: \"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3\") " Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.298759 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-utilities" (OuterVolumeSpecName: "utilities") pod "4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" (UID: "4954e3ec-e1f6-46dd-98b6-65ade22cc3c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.305008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-kube-api-access-n8jsh" (OuterVolumeSpecName: "kube-api-access-n8jsh") pod "4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" (UID: "4954e3ec-e1f6-46dd-98b6-65ade22cc3c3"). InnerVolumeSpecName "kube-api-access-n8jsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.355395 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" (UID: "4954e3ec-e1f6-46dd-98b6-65ade22cc3c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.399758 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.399792 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.399806 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8jsh\" (UniqueName: \"kubernetes.io/projected/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3-kube-api-access-n8jsh\") on node \"crc\" DevicePath \"\"" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.588022 4894 generic.go:334] "Generic (PLEG): container finished" podID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerID="69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e" exitCode=0 Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.588065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerDied","Data":"69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e"} Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.588094 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs977" event={"ID":"4954e3ec-e1f6-46dd-98b6-65ade22cc3c3","Type":"ContainerDied","Data":"2b2df8025be0c35ba0f53737c85ef5642c67915e37889090416fa6cceee1ff9c"} Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.588122 4894 scope.go:117] "RemoveContainer" containerID="69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.588297 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs977" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.624918 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs977"] Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.630028 4894 scope.go:117] "RemoveContainer" containerID="2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.635668 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs977"] Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.656028 4894 scope.go:117] "RemoveContainer" containerID="19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.705576 4894 scope.go:117] "RemoveContainer" containerID="69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e" Dec 10 19:50:03 crc kubenswrapper[4894]: E1210 19:50:03.710102 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e\": container with ID starting with 69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e not found: ID does not exist" containerID="69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.710147 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e"} err="failed to get container status \"69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e\": rpc error: code = NotFound desc = could not find container \"69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e\": container with ID starting with 69a8a00ef9a7f597e05a64ddac3b08256a505a9405c71a38d2a84dd74abd329e not found: ID does not exist" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.710177 4894 scope.go:117] "RemoveContainer" containerID="2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55" Dec 10 19:50:03 crc kubenswrapper[4894]: E1210 19:50:03.718504 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55\": container with ID starting with 2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55 not found: ID does not exist" containerID="2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.718561 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55"} err="failed to get container status \"2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55\": rpc error: code = NotFound desc = could not find container \"2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55\": container with ID starting with 2539fe09de12b3d8219f329572bac0ec7f3a08930bd3bae952305b81c6da5c55 not found: ID does not exist" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.718594 4894 scope.go:117] "RemoveContainer" containerID="19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12" Dec 10 19:50:03 crc kubenswrapper[4894]: E1210 19:50:03.719058 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12\": container with ID starting with 19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12 not found: ID does not exist" containerID="19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12" Dec 10 19:50:03 crc kubenswrapper[4894]: I1210 19:50:03.719124 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12"} err="failed to get container status \"19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12\": rpc error: code = NotFound desc = could not find container \"19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12\": container with ID starting with 19a3dff9289c5be4154a0c01f8a122b71118bab4475c03346df0b82608044e12 not found: ID does not exist" Dec 10 19:50:05 crc kubenswrapper[4894]: I1210 19:50:05.563544 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" path="/var/lib/kubelet/pods/4954e3ec-e1f6-46dd-98b6-65ade22cc3c3/volumes" Dec 10 19:50:14 crc kubenswrapper[4894]: I1210 19:50:14.540920 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:50:14 crc kubenswrapper[4894]: E1210 19:50:14.541719 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:50:29 crc kubenswrapper[4894]: I1210 19:50:29.540726 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:50:29 crc kubenswrapper[4894]: I1210 19:50:29.840865 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"c97814d79b586aec3e3def84ce058cc84c5bf15f31dad2edeae60269f9bb17eb"} Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.527293 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8f64j"] Dec 10 19:51:02 crc kubenswrapper[4894]: E1210 19:51:02.528252 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="extract-utilities" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.528269 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="extract-utilities" Dec 10 19:51:02 crc kubenswrapper[4894]: E1210 19:51:02.528306 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="registry-server" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.528314 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="registry-server" Dec 10 19:51:02 crc kubenswrapper[4894]: E1210 19:51:02.528338 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="extract-content" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.528345 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="extract-content" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.528590 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4954e3ec-e1f6-46dd-98b6-65ade22cc3c3" containerName="registry-server" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.530431 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.534648 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e3d76d-cb5c-43f8-aced-3981046a3daf-catalog-content\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.534714 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e3d76d-cb5c-43f8-aced-3981046a3daf-utilities\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.534779 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn559\" (UniqueName: \"kubernetes.io/projected/02e3d76d-cb5c-43f8-aced-3981046a3daf-kube-api-access-zn559\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.545988 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8f64j"] Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.637811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e3d76d-cb5c-43f8-aced-3981046a3daf-catalog-content\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.637953 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e3d76d-cb5c-43f8-aced-3981046a3daf-utilities\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.638015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn559\" (UniqueName: \"kubernetes.io/projected/02e3d76d-cb5c-43f8-aced-3981046a3daf-kube-api-access-zn559\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.638458 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e3d76d-cb5c-43f8-aced-3981046a3daf-catalog-content\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.638491 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e3d76d-cb5c-43f8-aced-3981046a3daf-utilities\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.672063 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn559\" (UniqueName: \"kubernetes.io/projected/02e3d76d-cb5c-43f8-aced-3981046a3daf-kube-api-access-zn559\") pod \"redhat-operators-8f64j\" (UID: \"02e3d76d-cb5c-43f8-aced-3981046a3daf\") " pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:02 crc kubenswrapper[4894]: I1210 19:51:02.856120 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:03 crc kubenswrapper[4894]: I1210 19:51:03.314176 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8f64j"] Dec 10 19:51:03 crc kubenswrapper[4894]: W1210 19:51:03.325028 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02e3d76d_cb5c_43f8_aced_3981046a3daf.slice/crio-8c2b5f0315b06053dd6a35c0bf2fbd046179fe053cc9e69bb5a02fbbdfb1eb3e WatchSource:0}: Error finding container 8c2b5f0315b06053dd6a35c0bf2fbd046179fe053cc9e69bb5a02fbbdfb1eb3e: Status 404 returned error can't find the container with id 8c2b5f0315b06053dd6a35c0bf2fbd046179fe053cc9e69bb5a02fbbdfb1eb3e Dec 10 19:51:04 crc kubenswrapper[4894]: I1210 19:51:04.185837 4894 generic.go:334] "Generic (PLEG): container finished" podID="02e3d76d-cb5c-43f8-aced-3981046a3daf" containerID="d42b48fc7d09a98fe1d899cf060edb81269827ff9dc8c035ce6c8208ba28257d" exitCode=0 Dec 10 19:51:04 crc kubenswrapper[4894]: I1210 19:51:04.185901 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8f64j" event={"ID":"02e3d76d-cb5c-43f8-aced-3981046a3daf","Type":"ContainerDied","Data":"d42b48fc7d09a98fe1d899cf060edb81269827ff9dc8c035ce6c8208ba28257d"} Dec 10 19:51:04 crc kubenswrapper[4894]: I1210 19:51:04.186166 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8f64j" event={"ID":"02e3d76d-cb5c-43f8-aced-3981046a3daf","Type":"ContainerStarted","Data":"8c2b5f0315b06053dd6a35c0bf2fbd046179fe053cc9e69bb5a02fbbdfb1eb3e"} Dec 10 19:51:15 crc kubenswrapper[4894]: I1210 19:51:15.315040 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8f64j" event={"ID":"02e3d76d-cb5c-43f8-aced-3981046a3daf","Type":"ContainerStarted","Data":"b13b55cbbc523c13c653d6c919a91bad96defff716b1c7675df68c3a4acc8c02"} Dec 10 19:51:18 crc kubenswrapper[4894]: I1210 19:51:18.352681 4894 generic.go:334] "Generic (PLEG): container finished" podID="02e3d76d-cb5c-43f8-aced-3981046a3daf" containerID="b13b55cbbc523c13c653d6c919a91bad96defff716b1c7675df68c3a4acc8c02" exitCode=0 Dec 10 19:51:18 crc kubenswrapper[4894]: I1210 19:51:18.352749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8f64j" event={"ID":"02e3d76d-cb5c-43f8-aced-3981046a3daf","Type":"ContainerDied","Data":"b13b55cbbc523c13c653d6c919a91bad96defff716b1c7675df68c3a4acc8c02"} Dec 10 19:51:20 crc kubenswrapper[4894]: I1210 19:51:20.378801 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8f64j" event={"ID":"02e3d76d-cb5c-43f8-aced-3981046a3daf","Type":"ContainerStarted","Data":"521f80a3b52a09fed8b188ad5541ba144463e8e30d9ccedbcc5f4841bc211036"} Dec 10 19:51:20 crc kubenswrapper[4894]: I1210 19:51:20.407939 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8f64j" podStartSLOduration=3.379576772 podStartE2EDuration="18.407914028s" podCreationTimestamp="2025-12-10 19:51:02 +0000 UTC" firstStartedPulling="2025-12-10 19:51:04.187823109 +0000 UTC m=+3346.982670901" lastFinishedPulling="2025-12-10 19:51:19.216160355 +0000 UTC m=+3362.011008157" observedRunningTime="2025-12-10 19:51:20.406870929 +0000 UTC m=+3363.201718741" watchObservedRunningTime="2025-12-10 19:51:20.407914028 +0000 UTC m=+3363.202761820" Dec 10 19:51:22 crc kubenswrapper[4894]: I1210 19:51:22.856790 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:22 crc kubenswrapper[4894]: I1210 19:51:22.858291 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:23 crc kubenswrapper[4894]: I1210 19:51:23.905516 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8f64j" podUID="02e3d76d-cb5c-43f8-aced-3981046a3daf" containerName="registry-server" probeResult="failure" output=< Dec 10 19:51:23 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 19:51:23 crc kubenswrapper[4894]: > Dec 10 19:51:32 crc kubenswrapper[4894]: I1210 19:51:32.905712 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:32 crc kubenswrapper[4894]: I1210 19:51:32.961411 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8f64j" Dec 10 19:51:33 crc kubenswrapper[4894]: I1210 19:51:33.604141 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8f64j"] Dec 10 19:51:33 crc kubenswrapper[4894]: I1210 19:51:33.729369 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t98mg"] Dec 10 19:51:33 crc kubenswrapper[4894]: I1210 19:51:33.729635 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t98mg" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="registry-server" containerID="cri-o://53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9" gracePeriod=2 Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.383631 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.475687 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-catalog-content\") pod \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.475745 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-utilities\") pod \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.475893 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlr7w\" (UniqueName: \"kubernetes.io/projected/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-kube-api-access-vlr7w\") pod \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\" (UID: \"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa\") " Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.477549 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-utilities" (OuterVolumeSpecName: "utilities") pod "ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" (UID: "ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.488466 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-kube-api-access-vlr7w" (OuterVolumeSpecName: "kube-api-access-vlr7w") pod "ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" (UID: "ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa"). InnerVolumeSpecName "kube-api-access-vlr7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.538578 4894 generic.go:334] "Generic (PLEG): container finished" podID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerID="53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9" exitCode=0 Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.538645 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerDied","Data":"53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9"} Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.538669 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t98mg" event={"ID":"ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa","Type":"ContainerDied","Data":"a254811a8106633f4321a0ec56cd09d42d0e4c86391d1d39f1ea95a44877d40d"} Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.538715 4894 scope.go:117] "RemoveContainer" containerID="53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.538705 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t98mg" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.540446 4894 generic.go:334] "Generic (PLEG): container finished" podID="1239b753-aad1-41b8-8df4-62df779823e5" containerID="9ae27fdf3ada8622de58120a31915c881237dc5b11abb61fd6d3e2d7792d95f8" exitCode=0 Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.540735 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1239b753-aad1-41b8-8df4-62df779823e5","Type":"ContainerDied","Data":"9ae27fdf3ada8622de58120a31915c881237dc5b11abb61fd6d3e2d7792d95f8"} Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.563574 4894 scope.go:117] "RemoveContainer" containerID="a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.578394 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.578423 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlr7w\" (UniqueName: \"kubernetes.io/projected/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-kube-api-access-vlr7w\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.586025 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" (UID: "ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.600610 4894 scope.go:117] "RemoveContainer" containerID="6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.638753 4894 scope.go:117] "RemoveContainer" containerID="53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9" Dec 10 19:51:34 crc kubenswrapper[4894]: E1210 19:51:34.639274 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9\": container with ID starting with 53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9 not found: ID does not exist" containerID="53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.639320 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9"} err="failed to get container status \"53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9\": rpc error: code = NotFound desc = could not find container \"53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9\": container with ID starting with 53f83714774450b84fd902966d637206cd6ea28393b4741b3f6c9b7898331fe9 not found: ID does not exist" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.639352 4894 scope.go:117] "RemoveContainer" containerID="a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c" Dec 10 19:51:34 crc kubenswrapper[4894]: E1210 19:51:34.639697 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c\": container with ID starting with a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c not found: ID does not exist" containerID="a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.639719 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c"} err="failed to get container status \"a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c\": rpc error: code = NotFound desc = could not find container \"a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c\": container with ID starting with a0c4d44ca948a83075b0d329fed0694d8eaf084f0a650281c8a99c05919b947c not found: ID does not exist" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.639733 4894 scope.go:117] "RemoveContainer" containerID="6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0" Dec 10 19:51:34 crc kubenswrapper[4894]: E1210 19:51:34.641076 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0\": container with ID starting with 6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0 not found: ID does not exist" containerID="6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.641123 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0"} err="failed to get container status \"6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0\": rpc error: code = NotFound desc = could not find container \"6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0\": container with ID starting with 6bc7584d431405939e0f9f014a52585700edb9873b137e41c0c5d925b0e8abe0 not found: ID does not exist" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.680756 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.876672 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t98mg"] Dec 10 19:51:34 crc kubenswrapper[4894]: I1210 19:51:34.885571 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t98mg"] Dec 10 19:51:35 crc kubenswrapper[4894]: I1210 19:51:35.556956 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" path="/var/lib/kubelet/pods/ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa/volumes" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.082438 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.205761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.205882 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.205919 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ca-certs\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.205966 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhj5w\" (UniqueName: \"kubernetes.io/projected/1239b753-aad1-41b8-8df4-62df779823e5-kube-api-access-rhj5w\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.205982 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ssh-key\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.206011 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-temporary\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.206080 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-workdir\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.206098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-config-data\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.206133 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config-secret\") pod \"1239b753-aad1-41b8-8df4-62df779823e5\" (UID: \"1239b753-aad1-41b8-8df4-62df779823e5\") " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.207399 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-config-data" (OuterVolumeSpecName: "config-data") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.209609 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.211620 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1239b753-aad1-41b8-8df4-62df779823e5-kube-api-access-rhj5w" (OuterVolumeSpecName: "kube-api-access-rhj5w") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "kube-api-access-rhj5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.214793 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.242114 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.244669 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.268968 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.284076 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309071 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309112 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309124 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309133 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhj5w\" (UniqueName: \"kubernetes.io/projected/1239b753-aad1-41b8-8df4-62df779823e5-kube-api-access-rhj5w\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309144 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309152 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309161 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1239b753-aad1-41b8-8df4-62df779823e5-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.309170 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1239b753-aad1-41b8-8df4-62df779823e5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.335943 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.411246 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.568533 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1239b753-aad1-41b8-8df4-62df779823e5","Type":"ContainerDied","Data":"05855e15cfccb5fb0a438ccc888a3fcbbf260272de05f5e8a4ceec94e9bc392c"} Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.568944 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05855e15cfccb5fb0a438ccc888a3fcbbf260272de05f5e8a4ceec94e9bc392c" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.569022 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.608444 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1239b753-aad1-41b8-8df4-62df779823e5" (UID: "1239b753-aad1-41b8-8df4-62df779823e5"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:51:36 crc kubenswrapper[4894]: I1210 19:51:36.614763 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1239b753-aad1-41b8-8df4-62df779823e5-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.178049 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 19:51:41 crc kubenswrapper[4894]: E1210 19:51:41.179162 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="registry-server" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.179178 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="registry-server" Dec 10 19:51:41 crc kubenswrapper[4894]: E1210 19:51:41.179213 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="extract-content" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.179222 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="extract-content" Dec 10 19:51:41 crc kubenswrapper[4894]: E1210 19:51:41.179238 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1239b753-aad1-41b8-8df4-62df779823e5" containerName="tempest-tests-tempest-tests-runner" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.179250 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1239b753-aad1-41b8-8df4-62df779823e5" containerName="tempest-tests-tempest-tests-runner" Dec 10 19:51:41 crc kubenswrapper[4894]: E1210 19:51:41.179267 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="extract-utilities" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.179276 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="extract-utilities" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.179536 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1239b753-aad1-41b8-8df4-62df779823e5" containerName="tempest-tests-tempest-tests-runner" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.179575 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab8bbd33-0fe3-42a0-84ba-0a3c9705cbfa" containerName="registry-server" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.180779 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.187598 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-c8vz7" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.191120 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.319694 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.320086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4fsk\" (UniqueName: \"kubernetes.io/projected/0917a379-621c-48f8-927d-c72b434e72ef-kube-api-access-t4fsk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.421447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.421603 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4fsk\" (UniqueName: \"kubernetes.io/projected/0917a379-621c-48f8-927d-c72b434e72ef-kube-api-access-t4fsk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.422162 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.443014 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4fsk\" (UniqueName: \"kubernetes.io/projected/0917a379-621c-48f8-927d-c72b434e72ef-kube-api-access-t4fsk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.471979 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"0917a379-621c-48f8-927d-c72b434e72ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.518767 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 19:51:41 crc kubenswrapper[4894]: I1210 19:51:41.974630 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 19:51:42 crc kubenswrapper[4894]: I1210 19:51:42.634762 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"0917a379-621c-48f8-927d-c72b434e72ef","Type":"ContainerStarted","Data":"b4acc5be7202f2e8615396ebf37f539b69aeb4b761c835adac6e92b1f0b9999a"} Dec 10 19:51:44 crc kubenswrapper[4894]: I1210 19:51:44.654179 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"0917a379-621c-48f8-927d-c72b434e72ef","Type":"ContainerStarted","Data":"4444e34374b59bf5858c0d2406c5261d451d69b82f81983ebff4cffa6f5ed1ca"} Dec 10 19:51:44 crc kubenswrapper[4894]: I1210 19:51:44.670150 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.162650687 podStartE2EDuration="3.67013265s" podCreationTimestamp="2025-12-10 19:51:41 +0000 UTC" firstStartedPulling="2025-12-10 19:51:41.981178362 +0000 UTC m=+3384.776026154" lastFinishedPulling="2025-12-10 19:51:43.488660305 +0000 UTC m=+3386.283508117" observedRunningTime="2025-12-10 19:51:44.665413723 +0000 UTC m=+3387.460261545" watchObservedRunningTime="2025-12-10 19:51:44.67013265 +0000 UTC m=+3387.464980442" Dec 10 19:52:17 crc kubenswrapper[4894]: I1210 19:52:17.923234 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6k6k8/must-gather-8v4rn"] Dec 10 19:52:17 crc kubenswrapper[4894]: I1210 19:52:17.926349 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:17 crc kubenswrapper[4894]: I1210 19:52:17.929028 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6k6k8"/"default-dockercfg-z6lkc" Dec 10 19:52:17 crc kubenswrapper[4894]: I1210 19:52:17.930741 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6k6k8"/"openshift-service-ca.crt" Dec 10 19:52:17 crc kubenswrapper[4894]: I1210 19:52:17.938642 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6k6k8"/"kube-root-ca.crt" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.015835 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6k6k8/must-gather-8v4rn"] Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.061256 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d09cbc31-3d36-4863-9083-079bc4fa2d36-must-gather-output\") pod \"must-gather-8v4rn\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.061352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp4b6\" (UniqueName: \"kubernetes.io/projected/d09cbc31-3d36-4863-9083-079bc4fa2d36-kube-api-access-bp4b6\") pod \"must-gather-8v4rn\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.163632 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d09cbc31-3d36-4863-9083-079bc4fa2d36-must-gather-output\") pod \"must-gather-8v4rn\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.163689 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp4b6\" (UniqueName: \"kubernetes.io/projected/d09cbc31-3d36-4863-9083-079bc4fa2d36-kube-api-access-bp4b6\") pod \"must-gather-8v4rn\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.164347 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d09cbc31-3d36-4863-9083-079bc4fa2d36-must-gather-output\") pod \"must-gather-8v4rn\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.184534 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp4b6\" (UniqueName: \"kubernetes.io/projected/d09cbc31-3d36-4863-9083-079bc4fa2d36-kube-api-access-bp4b6\") pod \"must-gather-8v4rn\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.246368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:52:18 crc kubenswrapper[4894]: I1210 19:52:18.938643 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6k6k8/must-gather-8v4rn"] Dec 10 19:52:19 crc kubenswrapper[4894]: I1210 19:52:19.434728 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" event={"ID":"d09cbc31-3d36-4863-9083-079bc4fa2d36","Type":"ContainerStarted","Data":"d2e8f450864d08dfb133f4931d87f47d726251f1c53bb5ad93cb71c8bf0923a9"} Dec 10 19:52:25 crc kubenswrapper[4894]: I1210 19:52:25.504727 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" event={"ID":"d09cbc31-3d36-4863-9083-079bc4fa2d36","Type":"ContainerStarted","Data":"29bbf647d3092ab31d531650bdf0622bd17296a9861e12283e274ebc8fa13e2b"} Dec 10 19:52:26 crc kubenswrapper[4894]: I1210 19:52:26.515493 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" event={"ID":"d09cbc31-3d36-4863-9083-079bc4fa2d36","Type":"ContainerStarted","Data":"051cd872b6500d52e078187f73639ab1716bd68511e1d167003e3631da4b351c"} Dec 10 19:52:26 crc kubenswrapper[4894]: I1210 19:52:26.534980 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" podStartSLOduration=3.239935651 podStartE2EDuration="9.534959024s" podCreationTimestamp="2025-12-10 19:52:17 +0000 UTC" firstStartedPulling="2025-12-10 19:52:18.944541366 +0000 UTC m=+3421.739389158" lastFinishedPulling="2025-12-10 19:52:25.239564739 +0000 UTC m=+3428.034412531" observedRunningTime="2025-12-10 19:52:26.527084053 +0000 UTC m=+3429.321931845" watchObservedRunningTime="2025-12-10 19:52:26.534959024 +0000 UTC m=+3429.329806816" Dec 10 19:52:28 crc kubenswrapper[4894]: I1210 19:52:28.987397 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-nhtkb"] Dec 10 19:52:28 crc kubenswrapper[4894]: I1210 19:52:28.989177 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.170437 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6cq4\" (UniqueName: \"kubernetes.io/projected/16e8a2c1-594f-448d-8413-ddf9409d31f0-kube-api-access-f6cq4\") pod \"crc-debug-nhtkb\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.170730 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e8a2c1-594f-448d-8413-ddf9409d31f0-host\") pod \"crc-debug-nhtkb\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.273086 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6cq4\" (UniqueName: \"kubernetes.io/projected/16e8a2c1-594f-448d-8413-ddf9409d31f0-kube-api-access-f6cq4\") pod \"crc-debug-nhtkb\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.273137 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e8a2c1-594f-448d-8413-ddf9409d31f0-host\") pod \"crc-debug-nhtkb\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.273345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e8a2c1-594f-448d-8413-ddf9409d31f0-host\") pod \"crc-debug-nhtkb\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.294352 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6cq4\" (UniqueName: \"kubernetes.io/projected/16e8a2c1-594f-448d-8413-ddf9409d31f0-kube-api-access-f6cq4\") pod \"crc-debug-nhtkb\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.332113 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:52:29 crc kubenswrapper[4894]: I1210 19:52:29.552506 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" event={"ID":"16e8a2c1-594f-448d-8413-ddf9409d31f0","Type":"ContainerStarted","Data":"6e497e7e977769032510e87f3559a7c6ff0a814ce976f531a039f8e0726124e3"} Dec 10 19:52:41 crc kubenswrapper[4894]: I1210 19:52:41.699240 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" event={"ID":"16e8a2c1-594f-448d-8413-ddf9409d31f0","Type":"ContainerStarted","Data":"4d615c1233437f041694e8226ab452d557676e11273bcec99b7dac4ddc57cd11"} Dec 10 19:52:41 crc kubenswrapper[4894]: I1210 19:52:41.727516 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" podStartSLOduration=2.32612246 podStartE2EDuration="13.727478652s" podCreationTimestamp="2025-12-10 19:52:28 +0000 UTC" firstStartedPulling="2025-12-10 19:52:29.388093226 +0000 UTC m=+3432.182941008" lastFinishedPulling="2025-12-10 19:52:40.789449408 +0000 UTC m=+3443.584297200" observedRunningTime="2025-12-10 19:52:41.713238858 +0000 UTC m=+3444.508086660" watchObservedRunningTime="2025-12-10 19:52:41.727478652 +0000 UTC m=+3444.522326444" Dec 10 19:52:49 crc kubenswrapper[4894]: I1210 19:52:49.577087 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:52:49 crc kubenswrapper[4894]: I1210 19:52:49.577799 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:53:19 crc kubenswrapper[4894]: I1210 19:53:19.577539 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:53:19 crc kubenswrapper[4894]: I1210 19:53:19.578117 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:53:28 crc kubenswrapper[4894]: I1210 19:53:28.159957 4894 generic.go:334] "Generic (PLEG): container finished" podID="16e8a2c1-594f-448d-8413-ddf9409d31f0" containerID="4d615c1233437f041694e8226ab452d557676e11273bcec99b7dac4ddc57cd11" exitCode=0 Dec 10 19:53:28 crc kubenswrapper[4894]: I1210 19:53:28.160032 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" event={"ID":"16e8a2c1-594f-448d-8413-ddf9409d31f0","Type":"ContainerDied","Data":"4d615c1233437f041694e8226ab452d557676e11273bcec99b7dac4ddc57cd11"} Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.301222 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.335840 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-nhtkb"] Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.338998 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e8a2c1-594f-448d-8413-ddf9409d31f0-host\") pod \"16e8a2c1-594f-448d-8413-ddf9409d31f0\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.339140 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/16e8a2c1-594f-448d-8413-ddf9409d31f0-host" (OuterVolumeSpecName: "host") pod "16e8a2c1-594f-448d-8413-ddf9409d31f0" (UID: "16e8a2c1-594f-448d-8413-ddf9409d31f0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.339216 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6cq4\" (UniqueName: \"kubernetes.io/projected/16e8a2c1-594f-448d-8413-ddf9409d31f0-kube-api-access-f6cq4\") pod \"16e8a2c1-594f-448d-8413-ddf9409d31f0\" (UID: \"16e8a2c1-594f-448d-8413-ddf9409d31f0\") " Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.339898 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/16e8a2c1-594f-448d-8413-ddf9409d31f0-host\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.345786 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16e8a2c1-594f-448d-8413-ddf9409d31f0-kube-api-access-f6cq4" (OuterVolumeSpecName: "kube-api-access-f6cq4") pod "16e8a2c1-594f-448d-8413-ddf9409d31f0" (UID: "16e8a2c1-594f-448d-8413-ddf9409d31f0"). InnerVolumeSpecName "kube-api-access-f6cq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.346938 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-nhtkb"] Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.441753 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6cq4\" (UniqueName: \"kubernetes.io/projected/16e8a2c1-594f-448d-8413-ddf9409d31f0-kube-api-access-f6cq4\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:29 crc kubenswrapper[4894]: I1210 19:53:29.559811 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16e8a2c1-594f-448d-8413-ddf9409d31f0" path="/var/lib/kubelet/pods/16e8a2c1-594f-448d-8413-ddf9409d31f0/volumes" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.186410 4894 scope.go:117] "RemoveContainer" containerID="4d615c1233437f041694e8226ab452d557676e11273bcec99b7dac4ddc57cd11" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.186450 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-nhtkb" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.509030 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-jwmx7"] Dec 10 19:53:30 crc kubenswrapper[4894]: E1210 19:53:30.510166 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16e8a2c1-594f-448d-8413-ddf9409d31f0" containerName="container-00" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.510250 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="16e8a2c1-594f-448d-8413-ddf9409d31f0" containerName="container-00" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.510533 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="16e8a2c1-594f-448d-8413-ddf9409d31f0" containerName="container-00" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.511372 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.564890 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6qfn\" (UniqueName: \"kubernetes.io/projected/0f2262d5-d850-4b87-af32-5f27643fa80b-kube-api-access-c6qfn\") pod \"crc-debug-jwmx7\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.565301 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f2262d5-d850-4b87-af32-5f27643fa80b-host\") pod \"crc-debug-jwmx7\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.667491 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6qfn\" (UniqueName: \"kubernetes.io/projected/0f2262d5-d850-4b87-af32-5f27643fa80b-kube-api-access-c6qfn\") pod \"crc-debug-jwmx7\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.667799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f2262d5-d850-4b87-af32-5f27643fa80b-host\") pod \"crc-debug-jwmx7\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.667983 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f2262d5-d850-4b87-af32-5f27643fa80b-host\") pod \"crc-debug-jwmx7\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.685684 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6qfn\" (UniqueName: \"kubernetes.io/projected/0f2262d5-d850-4b87-af32-5f27643fa80b-kube-api-access-c6qfn\") pod \"crc-debug-jwmx7\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:30 crc kubenswrapper[4894]: I1210 19:53:30.840178 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:31 crc kubenswrapper[4894]: I1210 19:53:31.197254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" event={"ID":"0f2262d5-d850-4b87-af32-5f27643fa80b","Type":"ContainerStarted","Data":"25e86c73a85ff8c7def8e48a1fc5df2ed6c2dc97e63d3163b49fac59051aa479"} Dec 10 19:53:31 crc kubenswrapper[4894]: I1210 19:53:31.197551 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" event={"ID":"0f2262d5-d850-4b87-af32-5f27643fa80b","Type":"ContainerStarted","Data":"300593ce17b2b41e9170054f2e5b0daa786dd17d4ef1465907b55b61a429cacb"} Dec 10 19:53:31 crc kubenswrapper[4894]: I1210 19:53:31.211805 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" podStartSLOduration=1.211786112 podStartE2EDuration="1.211786112s" podCreationTimestamp="2025-12-10 19:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:53:31.207415754 +0000 UTC m=+3494.002263566" watchObservedRunningTime="2025-12-10 19:53:31.211786112 +0000 UTC m=+3494.006633904" Dec 10 19:53:32 crc kubenswrapper[4894]: I1210 19:53:32.207608 4894 generic.go:334] "Generic (PLEG): container finished" podID="0f2262d5-d850-4b87-af32-5f27643fa80b" containerID="25e86c73a85ff8c7def8e48a1fc5df2ed6c2dc97e63d3163b49fac59051aa479" exitCode=0 Dec 10 19:53:32 crc kubenswrapper[4894]: I1210 19:53:32.207966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" event={"ID":"0f2262d5-d850-4b87-af32-5f27643fa80b","Type":"ContainerDied","Data":"25e86c73a85ff8c7def8e48a1fc5df2ed6c2dc97e63d3163b49fac59051aa479"} Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.346502 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.385516 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-jwmx7"] Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.394464 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-jwmx7"] Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.426243 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f2262d5-d850-4b87-af32-5f27643fa80b-host\") pod \"0f2262d5-d850-4b87-af32-5f27643fa80b\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.426321 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6qfn\" (UniqueName: \"kubernetes.io/projected/0f2262d5-d850-4b87-af32-5f27643fa80b-kube-api-access-c6qfn\") pod \"0f2262d5-d850-4b87-af32-5f27643fa80b\" (UID: \"0f2262d5-d850-4b87-af32-5f27643fa80b\") " Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.426388 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f2262d5-d850-4b87-af32-5f27643fa80b-host" (OuterVolumeSpecName: "host") pod "0f2262d5-d850-4b87-af32-5f27643fa80b" (UID: "0f2262d5-d850-4b87-af32-5f27643fa80b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.426888 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f2262d5-d850-4b87-af32-5f27643fa80b-host\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.432500 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2262d5-d850-4b87-af32-5f27643fa80b-kube-api-access-c6qfn" (OuterVolumeSpecName: "kube-api-access-c6qfn") pod "0f2262d5-d850-4b87-af32-5f27643fa80b" (UID: "0f2262d5-d850-4b87-af32-5f27643fa80b"). InnerVolumeSpecName "kube-api-access-c6qfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.528820 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6qfn\" (UniqueName: \"kubernetes.io/projected/0f2262d5-d850-4b87-af32-5f27643fa80b-kube-api-access-c6qfn\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:33 crc kubenswrapper[4894]: I1210 19:53:33.558318 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2262d5-d850-4b87-af32-5f27643fa80b" path="/var/lib/kubelet/pods/0f2262d5-d850-4b87-af32-5f27643fa80b/volumes" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.228893 4894 scope.go:117] "RemoveContainer" containerID="25e86c73a85ff8c7def8e48a1fc5df2ed6c2dc97e63d3163b49fac59051aa479" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.228950 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-jwmx7" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.577655 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-f6rt6"] Dec 10 19:53:34 crc kubenswrapper[4894]: E1210 19:53:34.578369 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2262d5-d850-4b87-af32-5f27643fa80b" containerName="container-00" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.578381 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2262d5-d850-4b87-af32-5f27643fa80b" containerName="container-00" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.578579 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2262d5-d850-4b87-af32-5f27643fa80b" containerName="container-00" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.579306 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.648580 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-host\") pod \"crc-debug-f6rt6\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.649069 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh99c\" (UniqueName: \"kubernetes.io/projected/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-kube-api-access-xh99c\") pod \"crc-debug-f6rt6\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.751192 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh99c\" (UniqueName: \"kubernetes.io/projected/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-kube-api-access-xh99c\") pod \"crc-debug-f6rt6\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.751331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-host\") pod \"crc-debug-f6rt6\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.751459 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-host\") pod \"crc-debug-f6rt6\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.768569 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh99c\" (UniqueName: \"kubernetes.io/projected/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-kube-api-access-xh99c\") pod \"crc-debug-f6rt6\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: I1210 19:53:34.904759 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:34 crc kubenswrapper[4894]: W1210 19:53:34.935506 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50cf197a_c9e5_41c9_989b_5776ecbb6f7e.slice/crio-99690dca305fadf28feb420cb29218c354ef443af9612ee7d9d3c429cfb16dd7 WatchSource:0}: Error finding container 99690dca305fadf28feb420cb29218c354ef443af9612ee7d9d3c429cfb16dd7: Status 404 returned error can't find the container with id 99690dca305fadf28feb420cb29218c354ef443af9612ee7d9d3c429cfb16dd7 Dec 10 19:53:35 crc kubenswrapper[4894]: I1210 19:53:35.240476 4894 generic.go:334] "Generic (PLEG): container finished" podID="50cf197a-c9e5-41c9-989b-5776ecbb6f7e" containerID="fccfa4187eaece2f43950c0e11983ea13e27659de18d1fcd2a3ab0295621192b" exitCode=0 Dec 10 19:53:35 crc kubenswrapper[4894]: I1210 19:53:35.240542 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" event={"ID":"50cf197a-c9e5-41c9-989b-5776ecbb6f7e","Type":"ContainerDied","Data":"fccfa4187eaece2f43950c0e11983ea13e27659de18d1fcd2a3ab0295621192b"} Dec 10 19:53:35 crc kubenswrapper[4894]: I1210 19:53:35.241330 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" event={"ID":"50cf197a-c9e5-41c9-989b-5776ecbb6f7e","Type":"ContainerStarted","Data":"99690dca305fadf28feb420cb29218c354ef443af9612ee7d9d3c429cfb16dd7"} Dec 10 19:53:35 crc kubenswrapper[4894]: I1210 19:53:35.281095 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-f6rt6"] Dec 10 19:53:35 crc kubenswrapper[4894]: I1210 19:53:35.292297 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6k6k8/crc-debug-f6rt6"] Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.357538 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.380160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh99c\" (UniqueName: \"kubernetes.io/projected/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-kube-api-access-xh99c\") pod \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.380257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-host\") pod \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\" (UID: \"50cf197a-c9e5-41c9-989b-5776ecbb6f7e\") " Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.380348 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-host" (OuterVolumeSpecName: "host") pod "50cf197a-c9e5-41c9-989b-5776ecbb6f7e" (UID: "50cf197a-c9e5-41c9-989b-5776ecbb6f7e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.380997 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-host\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.387833 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-kube-api-access-xh99c" (OuterVolumeSpecName: "kube-api-access-xh99c") pod "50cf197a-c9e5-41c9-989b-5776ecbb6f7e" (UID: "50cf197a-c9e5-41c9-989b-5776ecbb6f7e"). InnerVolumeSpecName "kube-api-access-xh99c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:53:36 crc kubenswrapper[4894]: I1210 19:53:36.483249 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh99c\" (UniqueName: \"kubernetes.io/projected/50cf197a-c9e5-41c9-989b-5776ecbb6f7e-kube-api-access-xh99c\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:37 crc kubenswrapper[4894]: I1210 19:53:37.261914 4894 scope.go:117] "RemoveContainer" containerID="fccfa4187eaece2f43950c0e11983ea13e27659de18d1fcd2a3ab0295621192b" Dec 10 19:53:37 crc kubenswrapper[4894]: I1210 19:53:37.261982 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/crc-debug-f6rt6" Dec 10 19:53:37 crc kubenswrapper[4894]: I1210 19:53:37.564108 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50cf197a-c9e5-41c9-989b-5776ecbb6f7e" path="/var/lib/kubelet/pods/50cf197a-c9e5-41c9-989b-5776ecbb6f7e/volumes" Dec 10 19:53:49 crc kubenswrapper[4894]: I1210 19:53:49.576940 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:53:49 crc kubenswrapper[4894]: I1210 19:53:49.577747 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:53:49 crc kubenswrapper[4894]: I1210 19:53:49.577815 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:53:49 crc kubenswrapper[4894]: I1210 19:53:49.579059 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c97814d79b586aec3e3def84ce058cc84c5bf15f31dad2edeae60269f9bb17eb"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:53:49 crc kubenswrapper[4894]: I1210 19:53:49.579164 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://c97814d79b586aec3e3def84ce058cc84c5bf15f31dad2edeae60269f9bb17eb" gracePeriod=600 Dec 10 19:53:50 crc kubenswrapper[4894]: I1210 19:53:50.401281 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="c97814d79b586aec3e3def84ce058cc84c5bf15f31dad2edeae60269f9bb17eb" exitCode=0 Dec 10 19:53:50 crc kubenswrapper[4894]: I1210 19:53:50.401338 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"c97814d79b586aec3e3def84ce058cc84c5bf15f31dad2edeae60269f9bb17eb"} Dec 10 19:53:50 crc kubenswrapper[4894]: I1210 19:53:50.401680 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935"} Dec 10 19:53:50 crc kubenswrapper[4894]: I1210 19:53:50.401704 4894 scope.go:117] "RemoveContainer" containerID="aad53a1a410d406a7d41702e3bcfe9d9146b162272948cd55e938c76bed14772" Dec 10 19:53:59 crc kubenswrapper[4894]: I1210 19:53:59.288169 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88520de0-eedd-40d7-ad33-ee752e8f2ff5/init-config-reloader/0.log" Dec 10 19:53:59 crc kubenswrapper[4894]: I1210 19:53:59.540966 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88520de0-eedd-40d7-ad33-ee752e8f2ff5/config-reloader/0.log" Dec 10 19:53:59 crc kubenswrapper[4894]: I1210 19:53:59.766741 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88520de0-eedd-40d7-ad33-ee752e8f2ff5/init-config-reloader/0.log" Dec 10 19:53:59 crc kubenswrapper[4894]: I1210 19:53:59.813402 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_88520de0-eedd-40d7-ad33-ee752e8f2ff5/alertmanager/0.log" Dec 10 19:53:59 crc kubenswrapper[4894]: I1210 19:53:59.939499 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54c578ccd6-gksvt_585f9d29-8845-45a9-9dd3-8fba7042a4e3/barbican-api/0.log" Dec 10 19:53:59 crc kubenswrapper[4894]: I1210 19:53:59.967979 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54c578ccd6-gksvt_585f9d29-8845-45a9-9dd3-8fba7042a4e3/barbican-api-log/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.054293 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f9794c7b8-gr6xz_30ab0905-d039-4f32-9740-3b7bffaab6ce/barbican-keystone-listener/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.254789 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f9794c7b8-gr6xz_30ab0905-d039-4f32-9740-3b7bffaab6ce/barbican-keystone-listener-log/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.271567 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cbb59f5d9-2cqct_25d1c4d9-024f-4f80-97e8-8b3c7fb726b0/barbican-worker/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.318073 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cbb59f5d9-2cqct_25d1c4d9-024f-4f80-97e8-8b3c7fb726b0/barbican-worker-log/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.502316 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nwd46_b1c310ef-2846-4536-9978-2cfd1c9a606c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.606606 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_10e3654c-3444-4fd9-9163-0b93a2089f15/ceilometer-central-agent/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.740253 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_10e3654c-3444-4fd9-9163-0b93a2089f15/proxy-httpd/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.773702 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_10e3654c-3444-4fd9-9163-0b93a2089f15/sg-core/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.786394 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_10e3654c-3444-4fd9-9163-0b93a2089f15/ceilometer-notification-agent/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.959120 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5ccf2774-6d3c-4b78-b027-867fbf8b9af0/cinder-api/0.log" Dec 10 19:54:00 crc kubenswrapper[4894]: I1210 19:54:00.987522 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5ccf2774-6d3c-4b78-b027-867fbf8b9af0/cinder-api-log/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.065375 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_de91d530-228b-482b-baf4-b7b4b1dd6fec/cinder-scheduler/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.159683 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_de91d530-228b-482b-baf4-b7b4b1dd6fec/probe/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.286206 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_5e167ef7-40fe-48c3-92b8-670eff58b203/cloudkitty-api-log/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.302314 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_5e167ef7-40fe-48c3-92b8-670eff58b203/cloudkitty-api/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.420647 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_25bf8748-8765-4f93-8edf-c4755dfe89bb/loki-compactor/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.487097 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-66dfd9bb-4qx65_028272b8-eb6a-4981-86b0-9438be5c55d3/loki-distributor/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.659468 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-7db4f4db8c-75nc2_e3561537-bc07-4e04-8e80-7ff3b376d8c9/gateway/0.log" Dec 10 19:54:01 crc kubenswrapper[4894]: I1210 19:54:01.793962 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-7db4f4db8c-vd8kl_9f05b972-ebf9-4b08-95c0-224225ef4271/gateway/0.log" Dec 10 19:54:02 crc kubenswrapper[4894]: I1210 19:54:02.014795 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_d5987e81-6ef3-4d95-88dc-7c05d531075f/loki-index-gateway/0.log" Dec 10 19:54:02 crc kubenswrapper[4894]: I1210 19:54:02.251016 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_da36895a-2989-4110-88a2-6bf0eb5d2f5d/loki-ingester/0.log" Dec 10 19:54:02 crc kubenswrapper[4894]: I1210 19:54:02.441109 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-5cd44666df-n7r2h_09e53cf2-d31c-4270-9002-29d42f94ea00/loki-query-frontend/0.log" Dec 10 19:54:02 crc kubenswrapper[4894]: I1210 19:54:02.658467 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-795fd8f8cc-rhw5x_bb4a57d8-a817-4bc8-9ffa-7f151d44d9ae/loki-querier/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.133992 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6zjzr_fd75fa45-6291-40b2-b40b-8763e6e72229/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.139351 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-597ft_bc921db2-6d84-4c34-9de5-12966910f500/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.346728 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5475ccd585-g6mjv_505346a3-228e-4aa5-8380-b057187ca6d0/init/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.542297 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5475ccd585-g6mjv_505346a3-228e-4aa5-8380-b057187ca6d0/init/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.590673 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5475ccd585-g6mjv_505346a3-228e-4aa5-8380-b057187ca6d0/dnsmasq-dns/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.689022 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4bqjt_25aa478e-2e66-4ba8-9428-17ad67e64df0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.822113 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e485201c-3c59-4daa-ba4d-f7cb8e91c3ec/glance-log/0.log" Dec 10 19:54:03 crc kubenswrapper[4894]: I1210 19:54:03.823376 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e485201c-3c59-4daa-ba4d-f7cb8e91c3ec/glance-httpd/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.161073 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_46843b9e-90e7-4d1d-9ae6-3be077fadfe5/glance-log/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.171573 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_46843b9e-90e7-4d1d-9ae6-3be077fadfe5/glance-httpd/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.242763 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-556fh_1a8fb1ad-f36f-4295-bdd9-3a5fa8e784eb/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.389983 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-vh26m_2583f762-94a3-4dab-9ace-2e1933c0af69/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.651717 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f8a3e9db-6a93-400c-94b2-1d7c5790dc17/kube-state-metrics/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.887460 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5w5tf_a5f93c12-d3db-42db-b431-861c46535ba4/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.906675 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6bf6fc85f8-h5xlr_06f64e80-03ab-47da-8dd6-670aaa6f1a32/keystone-api/0.log" Dec 10 19:54:04 crc kubenswrapper[4894]: I1210 19:54:04.960210 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_26836cdd-cd71-4430-9a64-bfd4cfa982c6/cloudkitty-proc/0.log" Dec 10 19:54:05 crc kubenswrapper[4894]: I1210 19:54:05.306519 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54d58d7797-fnl7d_8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8/neutron-httpd/0.log" Dec 10 19:54:05 crc kubenswrapper[4894]: I1210 19:54:05.355896 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54d58d7797-fnl7d_8ec216f2-1c8f-4e81-a7a1-6418d8a07cc8/neutron-api/0.log" Dec 10 19:54:05 crc kubenswrapper[4894]: I1210 19:54:05.467730 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-q6qrg_a9368153-c2ba-42ac-b7a5-5d47b02d8d14/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.114438 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c10fac35-af08-4efc-b7ec-e7a39bdf0245/nova-api-log/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.192131 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e75c654f-e152-4544-b6bd-a93466dcb0fc/nova-cell0-conductor-conductor/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.247092 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c10fac35-af08-4efc-b7ec-e7a39bdf0245/nova-api-api/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.428376 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b5404726-0e64-47ba-b50e-07e392a4d737/nova-cell1-conductor-conductor/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.517800 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_921bff93-5bc2-41c2-9d44-eac7ccdc50c1/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.679887 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-4dzqq_31e2c494-b1f4-453b-a536-26ac016c63f9/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:06 crc kubenswrapper[4894]: I1210 19:54:06.823375 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e4c4f31f-885b-4386-b6ae-2a0eb625b8d4/nova-metadata-log/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.073479 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a21758b8-7fad-4aff-ab6e-131b65d205ab/nova-scheduler-scheduler/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.154520 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab75b963-c89e-4871-8cb1-180dbaaf2714/mysql-bootstrap/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.443397 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab75b963-c89e-4871-8cb1-180dbaaf2714/galera/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.468513 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab75b963-c89e-4871-8cb1-180dbaaf2714/mysql-bootstrap/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.669289 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0297c834-eee7-4414-ba18-b4ee6b95561f/mysql-bootstrap/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.782808 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e4c4f31f-885b-4386-b6ae-2a0eb625b8d4/nova-metadata-metadata/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.893918 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0297c834-eee7-4414-ba18-b4ee6b95561f/galera/0.log" Dec 10 19:54:07 crc kubenswrapper[4894]: I1210 19:54:07.896204 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0297c834-eee7-4414-ba18-b4ee6b95561f/mysql-bootstrap/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.006456 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3d4126ef-dbcf-4614-a1fe-37c7c0be7223/openstackclient/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.121435 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-h2552_5fd5a07b-caf3-44d6-a9d4-e50e5be1819c/ovn-controller/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.252434 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gl9nv_023ba063-9ce8-45f2-aa5e-9eb7ee33e79c/openstack-network-exporter/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.437546 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pdwjv_df6eb3cc-342e-4d18-8027-db2c751fbaab/ovsdb-server-init/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.846133 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pdwjv_df6eb3cc-342e-4d18-8027-db2c751fbaab/ovs-vswitchd/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.900321 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pdwjv_df6eb3cc-342e-4d18-8027-db2c751fbaab/ovsdb-server/0.log" Dec 10 19:54:08 crc kubenswrapper[4894]: I1210 19:54:08.921358 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pdwjv_df6eb3cc-342e-4d18-8027-db2c751fbaab/ovsdb-server-init/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.104308 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-fhwr7_c91b533a-4f71-4b78-9442-a6d7bf300bff/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.132111 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ec34ebe8-9b05-4fc8-9866-d6cb02e61504/openstack-network-exporter/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.151969 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ec34ebe8-9b05-4fc8-9866-d6cb02e61504/ovn-northd/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.358314 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f7ec72a7-3043-4044-adfc-7ca2bf0c1749/ovsdbserver-nb/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.392837 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f7ec72a7-3043-4044-adfc-7ca2bf0c1749/openstack-network-exporter/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.593400 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b/openstack-network-exporter/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.608593 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5f6b1038-e1c2-4c8e-b06a-9edc1d48e17b/ovsdbserver-sb/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.859650 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7ff967f45b-c9flb_6a3f07cf-68dd-4c05-9cdc-2e4dd6031070/placement-api/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.935913 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7ff967f45b-c9flb_6a3f07cf-68dd-4c05-9cdc-2e4dd6031070/placement-log/0.log" Dec 10 19:54:09 crc kubenswrapper[4894]: I1210 19:54:09.981407 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dae05eff-e7b1-460a-a374-74e698d8f0c9/init-config-reloader/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.183960 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dae05eff-e7b1-460a-a374-74e698d8f0c9/config-reloader/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.195885 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dae05eff-e7b1-460a-a374-74e698d8f0c9/init-config-reloader/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.210251 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dae05eff-e7b1-460a-a374-74e698d8f0c9/thanos-sidecar/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.226024 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dae05eff-e7b1-460a-a374-74e698d8f0c9/prometheus/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.396703 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_800b4501-b71a-4854-88f9-4651f92b375e/setup-container/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.603278 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_800b4501-b71a-4854-88f9-4651f92b375e/rabbitmq/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.667633 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_800b4501-b71a-4854-88f9-4651f92b375e/setup-container/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.730725 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3d27beb4-0b6e-4543-b3a7-153fc30e2fe4/setup-container/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.872595 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3d27beb4-0b6e-4543-b3a7-153fc30e2fe4/rabbitmq/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.892920 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3d27beb4-0b6e-4543-b3a7-153fc30e2fe4/setup-container/0.log" Dec 10 19:54:10 crc kubenswrapper[4894]: I1210 19:54:10.947290 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-rk59n_710e20ed-7ffe-4041-9220-8ffa6b12b565/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.076646 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-dwdzp_fc05ca8d-4580-44b9-aad0-b23ce366f7ab/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.176536 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lth47_919d6c2a-def8-4ef2-b1b0-9f65d61be6fe/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.379261 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-d4gh8_11c0d6dd-5b70-4fe8-b69e-a1849ee6e6e0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.428821 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-st6mh_94091bd6-87ce-46be-8410-f881ac9f0d66/ssh-known-hosts-edpm-deployment/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.693932 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-75b94f784c-g8jm2_7c4b6f49-6469-47df-9f7c-4810b5bde4f0/proxy-server/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.710461 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-75b94f784c-g8jm2_7c4b6f49-6469-47df-9f7c-4810b5bde4f0/proxy-httpd/0.log" Dec 10 19:54:11 crc kubenswrapper[4894]: I1210 19:54:11.883651 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-mxbcj_fee50bb9-aa22-44b0-b596-c3f64d22c8a3/swift-ring-rebalance/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.004591 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/account-reaper/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.019993 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/account-auditor/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.142040 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/account-replicator/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.157040 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/account-server/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.248222 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/container-replicator/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.256637 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/container-auditor/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.352837 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/container-updater/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.363583 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/container-server/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.477263 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/object-auditor/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.493944 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/object-expirer/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.571387 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/object-replicator/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.587341 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/object-server/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.711645 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/rsync/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.712873 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/object-updater/0.log" Dec 10 19:54:12 crc kubenswrapper[4894]: I1210 19:54:12.835557 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5eac36f4-3cbc-48c1-9c6a-4e79de593d04/swift-recon-cron/0.log" Dec 10 19:54:13 crc kubenswrapper[4894]: I1210 19:54:13.086510 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-bsm59_08cd9bee-4229-441c-9eb5-4d6beea60c79/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:13 crc kubenswrapper[4894]: I1210 19:54:13.116035 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1239b753-aad1-41b8-8df4-62df779823e5/tempest-tests-tempest-tests-runner/0.log" Dec 10 19:54:13 crc kubenswrapper[4894]: I1210 19:54:13.283297 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_0917a379-621c-48f8-927d-c72b434e72ef/test-operator-logs-container/0.log" Dec 10 19:54:13 crc kubenswrapper[4894]: I1210 19:54:13.338207 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ht74m_7a4c1129-0acd-4a3e-a258-281efc367831/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 19:54:15 crc kubenswrapper[4894]: I1210 19:54:15.900745 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c7f34d67-8b26-449d-b38f-524e352efdd2/memcached/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.442992 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/util/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.632547 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/util/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.640321 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/pull/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.654872 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/pull/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.797732 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/util/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.820136 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/pull/0.log" Dec 10 19:54:38 crc kubenswrapper[4894]: I1210 19:54:38.856474 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7f7zq9x_e180e484-98a3-4635-9b27-fdaa1687d3f6/extract/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.025592 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-99gdz_3a81ced4-fe81-41ca-9623-243b811b3491/kube-rbac-proxy/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.057263 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-99gdz_3a81ced4-fe81-41ca-9623-243b811b3491/manager/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.089517 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-kkcfl_2fd92e49-8ea8-44c1-b516-a47c2c9bc516/kube-rbac-proxy/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.261953 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-kkcfl_2fd92e49-8ea8-44c1-b516-a47c2c9bc516/manager/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.270089 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-c7qtj_badfd685-2701-4ced-a77e-6eb7b7ff8c01/kube-rbac-proxy/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.335753 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-c7qtj_badfd685-2701-4ced-a77e-6eb7b7ff8c01/manager/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.486732 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-xb72b_6a960bb0-9f2c-4610-a418-791038fd1a06/kube-rbac-proxy/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.589442 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-xb72b_6a960bb0-9f2c-4610-a418-791038fd1a06/manager/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.667374 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lq68m_1a65ee6c-f87e-4183-8c8c-9582347febda/kube-rbac-proxy/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.713036 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lq68m_1a65ee6c-f87e-4183-8c8c-9582347febda/manager/0.log" Dec 10 19:54:39 crc kubenswrapper[4894]: I1210 19:54:39.891130 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-6fvml_f0669fb0-6164-4e2a-8f38-94d3a288923e/kube-rbac-proxy/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.128454 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-6fvml_f0669fb0-6164-4e2a-8f38-94d3a288923e/manager/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.234115 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-wqv7b_119c37b8-34ea-403c-9d87-6f252ff9a983/kube-rbac-proxy/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.445876 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-6fk8l_df0a10a8-f92d-4cd9-843c-8769455c9b24/kube-rbac-proxy/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.517092 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-6fk8l_df0a10a8-f92d-4cd9-843c-8769455c9b24/manager/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.577754 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-wqv7b_119c37b8-34ea-403c-9d87-6f252ff9a983/manager/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.680309 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-wphkl_ea0450b1-401e-42af-9e33-f29b60af26ed/kube-rbac-proxy/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.797006 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-wphkl_ea0450b1-401e-42af-9e33-f29b60af26ed/manager/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.875037 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-jlzvg_1ca34247-7826-4547-9167-b6357ce63c62/kube-rbac-proxy/0.log" Dec 10 19:54:40 crc kubenswrapper[4894]: I1210 19:54:40.972717 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-jlzvg_1ca34247-7826-4547-9167-b6357ce63c62/manager/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.074516 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-kclgx_608e5221-67cf-4e69-9ae1-037b120ff932/kube-rbac-proxy/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.128283 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-kclgx_608e5221-67cf-4e69-9ae1-037b120ff932/manager/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.235109 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2j6cg_0f98c7f8-f73d-4ca7-8361-91a719ab74ae/kube-rbac-proxy/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.354715 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2j6cg_0f98c7f8-f73d-4ca7-8361-91a719ab74ae/manager/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.402190 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4dvmk_cc54976b-bc17-4cb5-8eb3-901a163c2ec8/kube-rbac-proxy/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.506215 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4dvmk_cc54976b-bc17-4cb5-8eb3-901a163c2ec8/manager/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.584994 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-qsjsk_e1ef63d6-b70d-4424-86a3-988318cb4241/kube-rbac-proxy/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.622560 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-qsjsk_e1ef63d6-b70d-4424-86a3-988318cb4241/manager/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.739719 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879frxskr_b3bce49e-40c1-4e29-9905-825760e0f21d/kube-rbac-proxy/0.log" Dec 10 19:54:41 crc kubenswrapper[4894]: I1210 19:54:41.756560 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879frxskr_b3bce49e-40c1-4e29-9905-825760e0f21d/manager/0.log" Dec 10 19:54:42 crc kubenswrapper[4894]: I1210 19:54:42.177807 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6b7b77566b-tszmr_1b3de11c-140a-427c-a958-a534599df8d9/operator/0.log" Dec 10 19:54:42 crc kubenswrapper[4894]: I1210 19:54:42.250712 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jz862_3a2552c3-74d8-4da0-aae9-12bb7b08a42b/registry-server/0.log" Dec 10 19:54:42 crc kubenswrapper[4894]: I1210 19:54:42.670715 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-cg4zq_1b2f13e5-e186-439b-8c83-fbdb422b567b/kube-rbac-proxy/0.log" Dec 10 19:54:42 crc kubenswrapper[4894]: I1210 19:54:42.793870 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-cg4zq_1b2f13e5-e186-439b-8c83-fbdb422b567b/manager/0.log" Dec 10 19:54:42 crc kubenswrapper[4894]: I1210 19:54:42.876099 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4dkpz_08615ae3-02b0-41fb-8951-433da5904be5/kube-rbac-proxy/0.log" Dec 10 19:54:42 crc kubenswrapper[4894]: I1210 19:54:42.994954 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-4dkpz_08615ae3-02b0-41fb-8951-433da5904be5/manager/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.070277 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mzm6b_e1fb825a-ebc5-4cdf-949c-83d9bfe6c5af/operator/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.098322 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7b5585cf46-x9dnx_5d7bf121-f20c-463e-b9f6-8cebc9c6e5dd/manager/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.219367 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-c96bw_ddfceacd-628c-4f2f-bdc5-2c8cd7329098/kube-rbac-proxy/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.275593 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-c96bw_ddfceacd-628c-4f2f-bdc5-2c8cd7329098/manager/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.313833 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-954d6f856-m8wb7_7dd9f4f8-92dc-496a-a9b1-647a1619b7a5/kube-rbac-proxy/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.508479 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rmkmh_74c8428d-0015-4712-8bfa-6bf12f68b0bf/kube-rbac-proxy/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.590831 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rmkmh_74c8428d-0015-4712-8bfa-6bf12f68b0bf/manager/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.770029 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-954d6f856-m8wb7_7dd9f4f8-92dc-496a-a9b1-647a1619b7a5/manager/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.783579 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-qlq67_bcd78868-eb7e-4dfd-8992-2dd463f60287/manager/0.log" Dec 10 19:54:43 crc kubenswrapper[4894]: I1210 19:54:43.793704 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-qlq67_bcd78868-eb7e-4dfd-8992-2dd463f60287/kube-rbac-proxy/0.log" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.214501 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6hhg2"] Dec 10 19:54:51 crc kubenswrapper[4894]: E1210 19:54:51.215492 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50cf197a-c9e5-41c9-989b-5776ecbb6f7e" containerName="container-00" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.215504 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="50cf197a-c9e5-41c9-989b-5776ecbb6f7e" containerName="container-00" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.215712 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="50cf197a-c9e5-41c9-989b-5776ecbb6f7e" containerName="container-00" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.217988 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.230574 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6hhg2"] Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.367191 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85t72\" (UniqueName: \"kubernetes.io/projected/335008be-61c9-4187-b818-4dc7cd76b273-kube-api-access-85t72\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.367492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-catalog-content\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.367654 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-utilities\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.469780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85t72\" (UniqueName: \"kubernetes.io/projected/335008be-61c9-4187-b818-4dc7cd76b273-kube-api-access-85t72\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.469949 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-catalog-content\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.470004 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-utilities\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.470616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-catalog-content\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.470658 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-utilities\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.499814 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85t72\" (UniqueName: \"kubernetes.io/projected/335008be-61c9-4187-b818-4dc7cd76b273-kube-api-access-85t72\") pod \"certified-operators-6hhg2\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:51 crc kubenswrapper[4894]: I1210 19:54:51.537704 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:54:52 crc kubenswrapper[4894]: I1210 19:54:52.056636 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6hhg2"] Dec 10 19:54:53 crc kubenswrapper[4894]: I1210 19:54:53.061015 4894 generic.go:334] "Generic (PLEG): container finished" podID="335008be-61c9-4187-b818-4dc7cd76b273" containerID="b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f" exitCode=0 Dec 10 19:54:53 crc kubenswrapper[4894]: I1210 19:54:53.061123 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerDied","Data":"b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f"} Dec 10 19:54:53 crc kubenswrapper[4894]: I1210 19:54:53.061302 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerStarted","Data":"9f923a6f107c36f6f003c67dfb62320955278fd595bf2f4b350d66493fffdd85"} Dec 10 19:54:53 crc kubenswrapper[4894]: I1210 19:54:53.063995 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:54:54 crc kubenswrapper[4894]: I1210 19:54:54.074150 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerStarted","Data":"7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036"} Dec 10 19:54:56 crc kubenswrapper[4894]: I1210 19:54:56.095694 4894 generic.go:334] "Generic (PLEG): container finished" podID="335008be-61c9-4187-b818-4dc7cd76b273" containerID="7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036" exitCode=0 Dec 10 19:54:56 crc kubenswrapper[4894]: I1210 19:54:56.095783 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerDied","Data":"7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036"} Dec 10 19:54:57 crc kubenswrapper[4894]: I1210 19:54:57.105818 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerStarted","Data":"c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b"} Dec 10 19:54:57 crc kubenswrapper[4894]: I1210 19:54:57.156818 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6hhg2" podStartSLOduration=2.7146775869999997 podStartE2EDuration="6.156796157s" podCreationTimestamp="2025-12-10 19:54:51 +0000 UTC" firstStartedPulling="2025-12-10 19:54:53.063650868 +0000 UTC m=+3575.858498660" lastFinishedPulling="2025-12-10 19:54:56.505769438 +0000 UTC m=+3579.300617230" observedRunningTime="2025-12-10 19:54:57.151003991 +0000 UTC m=+3579.945851783" watchObservedRunningTime="2025-12-10 19:54:57.156796157 +0000 UTC m=+3579.951643959" Dec 10 19:55:01 crc kubenswrapper[4894]: I1210 19:55:01.538234 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:55:01 crc kubenswrapper[4894]: I1210 19:55:01.538870 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:55:01 crc kubenswrapper[4894]: I1210 19:55:01.591182 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:55:02 crc kubenswrapper[4894]: I1210 19:55:02.197563 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:55:02 crc kubenswrapper[4894]: I1210 19:55:02.255451 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6hhg2"] Dec 10 19:55:03 crc kubenswrapper[4894]: I1210 19:55:03.334610 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-spk6j_a2b2f487-4291-4ed7-a585-2fb5ba3ed7e4/control-plane-machine-set-operator/0.log" Dec 10 19:55:03 crc kubenswrapper[4894]: I1210 19:55:03.560536 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bl258_85556679-8cbc-4321-a936-22924ac80223/kube-rbac-proxy/0.log" Dec 10 19:55:03 crc kubenswrapper[4894]: I1210 19:55:03.610121 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bl258_85556679-8cbc-4321-a936-22924ac80223/machine-api-operator/0.log" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.162048 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6hhg2" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="registry-server" containerID="cri-o://c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b" gracePeriod=2 Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.765788 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.878182 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-catalog-content\") pod \"335008be-61c9-4187-b818-4dc7cd76b273\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.878277 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85t72\" (UniqueName: \"kubernetes.io/projected/335008be-61c9-4187-b818-4dc7cd76b273-kube-api-access-85t72\") pod \"335008be-61c9-4187-b818-4dc7cd76b273\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.878380 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-utilities\") pod \"335008be-61c9-4187-b818-4dc7cd76b273\" (UID: \"335008be-61c9-4187-b818-4dc7cd76b273\") " Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.879192 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-utilities" (OuterVolumeSpecName: "utilities") pod "335008be-61c9-4187-b818-4dc7cd76b273" (UID: "335008be-61c9-4187-b818-4dc7cd76b273"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.887009 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/335008be-61c9-4187-b818-4dc7cd76b273-kube-api-access-85t72" (OuterVolumeSpecName: "kube-api-access-85t72") pod "335008be-61c9-4187-b818-4dc7cd76b273" (UID: "335008be-61c9-4187-b818-4dc7cd76b273"). InnerVolumeSpecName "kube-api-access-85t72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.925902 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "335008be-61c9-4187-b818-4dc7cd76b273" (UID: "335008be-61c9-4187-b818-4dc7cd76b273"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.981325 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85t72\" (UniqueName: \"kubernetes.io/projected/335008be-61c9-4187-b818-4dc7cd76b273-kube-api-access-85t72\") on node \"crc\" DevicePath \"\"" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.981356 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:55:04 crc kubenswrapper[4894]: I1210 19:55:04.981365 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335008be-61c9-4187-b818-4dc7cd76b273-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.175765 4894 generic.go:334] "Generic (PLEG): container finished" podID="335008be-61c9-4187-b818-4dc7cd76b273" containerID="c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b" exitCode=0 Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.175810 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerDied","Data":"c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b"} Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.175836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6hhg2" event={"ID":"335008be-61c9-4187-b818-4dc7cd76b273","Type":"ContainerDied","Data":"9f923a6f107c36f6f003c67dfb62320955278fd595bf2f4b350d66493fffdd85"} Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.175881 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6hhg2" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.175895 4894 scope.go:117] "RemoveContainer" containerID="c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.200643 4894 scope.go:117] "RemoveContainer" containerID="7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.229408 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6hhg2"] Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.238892 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6hhg2"] Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.251061 4894 scope.go:117] "RemoveContainer" containerID="b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.289041 4894 scope.go:117] "RemoveContainer" containerID="c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b" Dec 10 19:55:05 crc kubenswrapper[4894]: E1210 19:55:05.289517 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b\": container with ID starting with c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b not found: ID does not exist" containerID="c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.289560 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b"} err="failed to get container status \"c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b\": rpc error: code = NotFound desc = could not find container \"c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b\": container with ID starting with c99d199ac358a52c66ed1fc2b0fa24fe10d04957b9275c28af950fbc440b344b not found: ID does not exist" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.289587 4894 scope.go:117] "RemoveContainer" containerID="7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036" Dec 10 19:55:05 crc kubenswrapper[4894]: E1210 19:55:05.289981 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036\": container with ID starting with 7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036 not found: ID does not exist" containerID="7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.290010 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036"} err="failed to get container status \"7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036\": rpc error: code = NotFound desc = could not find container \"7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036\": container with ID starting with 7f6036ffa90568c7e19b3172cb54414086981bd09d62990201e4097210c91036 not found: ID does not exist" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.290031 4894 scope.go:117] "RemoveContainer" containerID="b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f" Dec 10 19:55:05 crc kubenswrapper[4894]: E1210 19:55:05.290430 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f\": container with ID starting with b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f not found: ID does not exist" containerID="b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.290462 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f"} err="failed to get container status \"b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f\": rpc error: code = NotFound desc = could not find container \"b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f\": container with ID starting with b9a41da238ffad4ab0564422b5f4d363a4e7bc3c4e04899580dd6c0563ca5e7f not found: ID does not exist" Dec 10 19:55:05 crc kubenswrapper[4894]: I1210 19:55:05.555290 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="335008be-61c9-4187-b818-4dc7cd76b273" path="/var/lib/kubelet/pods/335008be-61c9-4187-b818-4dc7cd76b273/volumes" Dec 10 19:55:16 crc kubenswrapper[4894]: I1210 19:55:16.501228 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vqs2f_451bafa9-45d0-4374-8bad-29067fd16d26/cert-manager-controller/0.log" Dec 10 19:55:16 crc kubenswrapper[4894]: I1210 19:55:16.616871 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qv9lh_1134d9d0-0ca0-4a51-8066-acc409fd3218/cert-manager-cainjector/0.log" Dec 10 19:55:16 crc kubenswrapper[4894]: I1210 19:55:16.662989 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-fsvcg_ca50f7b6-2bd0-48ce-af0c-9edf0b895058/cert-manager-webhook/0.log" Dec 10 19:55:29 crc kubenswrapper[4894]: I1210 19:55:29.708612 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-qcks8_5d0c2c6f-f293-48cd-a22b-d77fc2fd9b02/nmstate-console-plugin/0.log" Dec 10 19:55:29 crc kubenswrapper[4894]: I1210 19:55:29.921786 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-knvqx_f3832f69-624a-4be5-a69a-75e1d3200af4/nmstate-handler/0.log" Dec 10 19:55:29 crc kubenswrapper[4894]: I1210 19:55:29.954638 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-g868p_ab35ae17-b6ef-49e6-8cdb-625e03d9890d/kube-rbac-proxy/0.log" Dec 10 19:55:29 crc kubenswrapper[4894]: I1210 19:55:29.994216 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-g868p_ab35ae17-b6ef-49e6-8cdb-625e03d9890d/nmstate-metrics/0.log" Dec 10 19:55:30 crc kubenswrapper[4894]: I1210 19:55:30.174799 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-rh2pw_be4369f3-8902-40ee-9508-45698d96d6af/nmstate-webhook/0.log" Dec 10 19:55:30 crc kubenswrapper[4894]: I1210 19:55:30.177146 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-t5pwh_c7d3b102-8456-4801-944d-7fb2b90cafc4/nmstate-operator/0.log" Dec 10 19:55:42 crc kubenswrapper[4894]: I1210 19:55:42.567893 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-59fd4b8cb8-vkdlk_c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec/kube-rbac-proxy/0.log" Dec 10 19:55:42 crc kubenswrapper[4894]: I1210 19:55:42.616409 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-59fd4b8cb8-vkdlk_c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec/manager/0.log" Dec 10 19:55:49 crc kubenswrapper[4894]: I1210 19:55:49.576901 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:55:49 crc kubenswrapper[4894]: I1210 19:55:49.578397 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.375111 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-q2tvc_eb70538a-122e-480f-8853-15e96afdf9a4/kube-rbac-proxy/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.471674 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-q2tvc_eb70538a-122e-480f-8853-15e96afdf9a4/controller/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.553210 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-frr-files/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.726607 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-frr-files/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.753830 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-reloader/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.762494 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-metrics/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.767705 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-reloader/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.896747 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-frr-files/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.946760 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-metrics/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.957989 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-reloader/0.log" Dec 10 19:55:55 crc kubenswrapper[4894]: I1210 19:55:55.994468 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-metrics/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.132728 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-frr-files/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.154154 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-reloader/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.157648 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/controller/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.161451 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/cp-metrics/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.324136 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/kube-rbac-proxy/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.334339 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/kube-rbac-proxy-frr/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.354093 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/frr-metrics/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.503575 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/reloader/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.733573 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-skk4g_a3852d90-39db-4a6e-b4cd-4d805d18a2b5/frr-k8s-webhook-server/0.log" Dec 10 19:55:56 crc kubenswrapper[4894]: I1210 19:55:56.989305 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7c64f56f84-7stp8_60aa603c-1807-48aa-80c0-fd56eedeff49/webhook-server/0.log" Dec 10 19:55:57 crc kubenswrapper[4894]: I1210 19:55:57.010559 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8484c557fc-rnmp8_25e24603-77a9-43cd-86c6-941b45ab2d4b/manager/0.log" Dec 10 19:55:57 crc kubenswrapper[4894]: I1210 19:55:57.282042 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fpgb2_ffb10c89-1b21-48b2-8d11-320ccf17dd64/kube-rbac-proxy/0.log" Dec 10 19:55:57 crc kubenswrapper[4894]: I1210 19:55:57.774336 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fpgb2_ffb10c89-1b21-48b2-8d11-320ccf17dd64/speaker/0.log" Dec 10 19:55:57 crc kubenswrapper[4894]: I1210 19:55:57.953708 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-svgvf_3731d1e5-d0ee-4ad6-97f9-642fc2d0a3ee/frr/0.log" Dec 10 19:56:10 crc kubenswrapper[4894]: I1210 19:56:10.903377 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/util/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.096757 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/pull/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.099072 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/pull/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.111041 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/util/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.315047 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/pull/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.324720 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/extract/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.328435 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_3e572a74f8b8ca2bcfe04329d4f26bd9689911be5d166a7403bd6ae77388dk9_6fa7084b-6307-4bbb-b690-076cc71fb40c/util/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.491280 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/util/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.702532 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/util/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.737027 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/pull/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.772739 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/pull/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.905504 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/util/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.919748 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/pull/0.log" Dec 10 19:56:11 crc kubenswrapper[4894]: I1210 19:56:11.977677 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fxhqb4_877cd716-19e7-461e-b218-711dafd4f785/extract/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.096414 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/util/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.278956 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/util/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.331710 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/pull/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.361043 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/pull/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.654113 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/util/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.662674 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/pull/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.711127 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vtkjm_d191af44-bc04-4529-93a4-b58a773920f5/extract/0.log" Dec 10 19:56:12 crc kubenswrapper[4894]: I1210 19:56:12.871277 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/util/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.035941 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/pull/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.039336 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/util/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.069257 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/pull/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.244329 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/extract/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.246285 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/util/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.257764 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83p7vh2_e20fd98f-3383-4462-9894-ab6c867c0064/pull/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.426245 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/extract-utilities/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.581568 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/extract-utilities/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.584895 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/extract-content/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.589723 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/extract-content/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.822886 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/extract-utilities/0.log" Dec 10 19:56:13 crc kubenswrapper[4894]: I1210 19:56:13.844704 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/extract-content/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.017509 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/extract-utilities/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.268139 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/extract-utilities/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.287563 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/extract-content/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.365996 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/extract-content/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.398828 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v4lrz_6306ae09-754d-4285-8182-4f6433405855/registry-server/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.548995 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/extract-utilities/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.555762 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/extract-content/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.750443 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sgnhf_8c9b37fd-63d5-4b44-9e8e-d41b6ec233ea/marketplace-operator/0.log" Dec 10 19:56:14 crc kubenswrapper[4894]: I1210 19:56:14.900298 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/extract-utilities/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.054019 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/extract-utilities/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.083729 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/extract-content/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.117591 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/extract-content/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.138647 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c944h_d42b609b-dba3-484f-b919-a33fc18e3385/registry-server/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.253242 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/extract-utilities/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.270726 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/extract-content/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.392718 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6zg8t_e8892c15-a6a1-45d4-adf1-84b6e132b411/registry-server/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.402692 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/extract-utilities/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.558137 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/extract-content/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.562063 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/extract-content/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.565777 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/extract-utilities/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.751429 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/extract-utilities/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.988706 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/registry-server/0.log" Dec 10 19:56:15 crc kubenswrapper[4894]: I1210 19:56:15.989453 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-8f64j_02e3d76d-cb5c-43f8-aced-3981046a3daf/extract-content/0.log" Dec 10 19:56:19 crc kubenswrapper[4894]: I1210 19:56:19.576984 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:56:19 crc kubenswrapper[4894]: I1210 19:56:19.577677 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:56:28 crc kubenswrapper[4894]: I1210 19:56:28.244278 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-z6v6g_3d09cb3e-2427-4282-8112-806df77eb803/prometheus-operator/0.log" Dec 10 19:56:28 crc kubenswrapper[4894]: I1210 19:56:28.410182 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d79d4bb48-7xgxj_f23b2be9-d333-433e-8381-baec276a1b30/prometheus-operator-admission-webhook/0.log" Dec 10 19:56:28 crc kubenswrapper[4894]: I1210 19:56:28.452425 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6d79d4bb48-nzbzk_1b35687b-c053-4902-8ae4-146c82ae486c/prometheus-operator-admission-webhook/0.log" Dec 10 19:56:28 crc kubenswrapper[4894]: I1210 19:56:28.791312 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-89qkl_641b65b6-0781-479e-ac89-c1cdeb2e5b2d/operator/0.log" Dec 10 19:56:28 crc kubenswrapper[4894]: I1210 19:56:28.800230 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-67pkr_91e1df09-581a-4dc8-9dae-6834c2bb3bfe/perses-operator/0.log" Dec 10 19:56:41 crc kubenswrapper[4894]: I1210 19:56:41.575015 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-59fd4b8cb8-vkdlk_c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec/kube-rbac-proxy/0.log" Dec 10 19:56:41 crc kubenswrapper[4894]: I1210 19:56:41.627631 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-59fd4b8cb8-vkdlk_c37c56d9-ddfa-49ed-bebe-5ea24d31e9ec/manager/0.log" Dec 10 19:56:49 crc kubenswrapper[4894]: I1210 19:56:49.577511 4894 patch_prober.go:28] interesting pod/machine-config-daemon-vbjj4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:56:49 crc kubenswrapper[4894]: I1210 19:56:49.578231 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:56:49 crc kubenswrapper[4894]: I1210 19:56:49.578282 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" Dec 10 19:56:49 crc kubenswrapper[4894]: I1210 19:56:49.579173 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935"} pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:56:49 crc kubenswrapper[4894]: I1210 19:56:49.579240 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" containerName="machine-config-daemon" containerID="cri-o://088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" gracePeriod=600 Dec 10 19:56:49 crc kubenswrapper[4894]: E1210 19:56:49.706603 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:56:50 crc kubenswrapper[4894]: I1210 19:56:50.226002 4894 generic.go:334] "Generic (PLEG): container finished" podID="3fd55194-d263-4547-9b8f-934ef8050b59" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" exitCode=0 Dec 10 19:56:50 crc kubenswrapper[4894]: I1210 19:56:50.226097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerDied","Data":"088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935"} Dec 10 19:56:50 crc kubenswrapper[4894]: I1210 19:56:50.226418 4894 scope.go:117] "RemoveContainer" containerID="c97814d79b586aec3e3def84ce058cc84c5bf15f31dad2edeae60269f9bb17eb" Dec 10 19:56:50 crc kubenswrapper[4894]: I1210 19:56:50.227438 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:56:50 crc kubenswrapper[4894]: E1210 19:56:50.228037 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:57:01 crc kubenswrapper[4894]: I1210 19:57:01.541965 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:57:01 crc kubenswrapper[4894]: E1210 19:57:01.542862 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:57:12 crc kubenswrapper[4894]: I1210 19:57:12.541533 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:57:12 crc kubenswrapper[4894]: E1210 19:57:12.542470 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:57:25 crc kubenswrapper[4894]: I1210 19:57:25.541362 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:57:25 crc kubenswrapper[4894]: E1210 19:57:25.543486 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:57:39 crc kubenswrapper[4894]: I1210 19:57:39.541804 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:57:39 crc kubenswrapper[4894]: E1210 19:57:39.542782 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:57:51 crc kubenswrapper[4894]: I1210 19:57:51.542211 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:57:51 crc kubenswrapper[4894]: E1210 19:57:51.543009 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:58:04 crc kubenswrapper[4894]: I1210 19:58:04.542628 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:58:04 crc kubenswrapper[4894]: E1210 19:58:04.543624 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:58:16 crc kubenswrapper[4894]: I1210 19:58:16.542177 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:58:16 crc kubenswrapper[4894]: E1210 19:58:16.543195 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:58:22 crc kubenswrapper[4894]: I1210 19:58:22.255392 4894 generic.go:334] "Generic (PLEG): container finished" podID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerID="29bbf647d3092ab31d531650bdf0622bd17296a9861e12283e274ebc8fa13e2b" exitCode=0 Dec 10 19:58:22 crc kubenswrapper[4894]: I1210 19:58:22.255580 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" event={"ID":"d09cbc31-3d36-4863-9083-079bc4fa2d36","Type":"ContainerDied","Data":"29bbf647d3092ab31d531650bdf0622bd17296a9861e12283e274ebc8fa13e2b"} Dec 10 19:58:22 crc kubenswrapper[4894]: I1210 19:58:22.257106 4894 scope.go:117] "RemoveContainer" containerID="29bbf647d3092ab31d531650bdf0622bd17296a9861e12283e274ebc8fa13e2b" Dec 10 19:58:22 crc kubenswrapper[4894]: I1210 19:58:22.522264 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6k6k8_must-gather-8v4rn_d09cbc31-3d36-4863-9083-079bc4fa2d36/gather/0.log" Dec 10 19:58:30 crc kubenswrapper[4894]: I1210 19:58:30.540804 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:58:30 crc kubenswrapper[4894]: E1210 19:58:30.541751 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.036879 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6k6k8/must-gather-8v4rn"] Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.037453 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="copy" containerID="cri-o://051cd872b6500d52e078187f73639ab1716bd68511e1d167003e3631da4b351c" gracePeriod=2 Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.051034 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6k6k8/must-gather-8v4rn"] Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.359332 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6k6k8_must-gather-8v4rn_d09cbc31-3d36-4863-9083-079bc4fa2d36/copy/0.log" Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.360298 4894 generic.go:334] "Generic (PLEG): container finished" podID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerID="051cd872b6500d52e078187f73639ab1716bd68511e1d167003e3631da4b351c" exitCode=143 Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.610115 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6k6k8_must-gather-8v4rn_d09cbc31-3d36-4863-9083-079bc4fa2d36/copy/0.log" Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.610522 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.759653 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d09cbc31-3d36-4863-9083-079bc4fa2d36-must-gather-output\") pod \"d09cbc31-3d36-4863-9083-079bc4fa2d36\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.759840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp4b6\" (UniqueName: \"kubernetes.io/projected/d09cbc31-3d36-4863-9083-079bc4fa2d36-kube-api-access-bp4b6\") pod \"d09cbc31-3d36-4863-9083-079bc4fa2d36\" (UID: \"d09cbc31-3d36-4863-9083-079bc4fa2d36\") " Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.783995 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d09cbc31-3d36-4863-9083-079bc4fa2d36-kube-api-access-bp4b6" (OuterVolumeSpecName: "kube-api-access-bp4b6") pod "d09cbc31-3d36-4863-9083-079bc4fa2d36" (UID: "d09cbc31-3d36-4863-9083-079bc4fa2d36"). InnerVolumeSpecName "kube-api-access-bp4b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.862760 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp4b6\" (UniqueName: \"kubernetes.io/projected/d09cbc31-3d36-4863-9083-079bc4fa2d36-kube-api-access-bp4b6\") on node \"crc\" DevicePath \"\"" Dec 10 19:58:32 crc kubenswrapper[4894]: I1210 19:58:32.970639 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d09cbc31-3d36-4863-9083-079bc4fa2d36-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d09cbc31-3d36-4863-9083-079bc4fa2d36" (UID: "d09cbc31-3d36-4863-9083-079bc4fa2d36"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:58:33 crc kubenswrapper[4894]: I1210 19:58:33.066273 4894 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d09cbc31-3d36-4863-9083-079bc4fa2d36-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 19:58:33 crc kubenswrapper[4894]: I1210 19:58:33.393940 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6k6k8_must-gather-8v4rn_d09cbc31-3d36-4863-9083-079bc4fa2d36/copy/0.log" Dec 10 19:58:33 crc kubenswrapper[4894]: I1210 19:58:33.394482 4894 scope.go:117] "RemoveContainer" containerID="051cd872b6500d52e078187f73639ab1716bd68511e1d167003e3631da4b351c" Dec 10 19:58:33 crc kubenswrapper[4894]: I1210 19:58:33.394656 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6k6k8/must-gather-8v4rn" Dec 10 19:58:33 crc kubenswrapper[4894]: I1210 19:58:33.445443 4894 scope.go:117] "RemoveContainer" containerID="29bbf647d3092ab31d531650bdf0622bd17296a9861e12283e274ebc8fa13e2b" Dec 10 19:58:33 crc kubenswrapper[4894]: I1210 19:58:33.591396 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" path="/var/lib/kubelet/pods/d09cbc31-3d36-4863-9083-079bc4fa2d36/volumes" Dec 10 19:58:41 crc kubenswrapper[4894]: I1210 19:58:41.541953 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:58:41 crc kubenswrapper[4894]: E1210 19:58:41.543108 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:58:54 crc kubenswrapper[4894]: I1210 19:58:54.541746 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:58:54 crc kubenswrapper[4894]: E1210 19:58:54.542561 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:59:06 crc kubenswrapper[4894]: I1210 19:59:06.541472 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:59:06 crc kubenswrapper[4894]: E1210 19:59:06.542370 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:59:20 crc kubenswrapper[4894]: I1210 19:59:20.541192 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:59:20 crc kubenswrapper[4894]: E1210 19:59:20.542176 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:59:33 crc kubenswrapper[4894]: I1210 19:59:33.541121 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:59:33 crc kubenswrapper[4894]: E1210 19:59:33.542303 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:59:44 crc kubenswrapper[4894]: I1210 19:59:44.541449 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:59:44 crc kubenswrapper[4894]: E1210 19:59:44.542134 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 19:59:59 crc kubenswrapper[4894]: I1210 19:59:59.542088 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 19:59:59 crc kubenswrapper[4894]: E1210 19:59:59.542956 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.191478 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf"] Dec 10 20:00:00 crc kubenswrapper[4894]: E1210 20:00:00.191964 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="extract-content" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.191981 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="extract-content" Dec 10 20:00:00 crc kubenswrapper[4894]: E1210 20:00:00.191999 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="copy" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192006 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="copy" Dec 10 20:00:00 crc kubenswrapper[4894]: E1210 20:00:00.192020 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192026 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4894]: E1210 20:00:00.192063 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="extract-utilities" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192073 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="extract-utilities" Dec 10 20:00:00 crc kubenswrapper[4894]: E1210 20:00:00.192093 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="gather" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192100 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="gather" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192343 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="copy" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192366 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d09cbc31-3d36-4863-9083-079bc4fa2d36" containerName="gather" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.192375 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="335008be-61c9-4187-b818-4dc7cd76b273" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.193286 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.194928 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.195171 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.203019 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf"] Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.235822 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e14dd6-904b-45e5-884c-f35ba1581393-config-volume\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.235950 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m44b\" (UniqueName: \"kubernetes.io/projected/41e14dd6-904b-45e5-884c-f35ba1581393-kube-api-access-6m44b\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.236296 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e14dd6-904b-45e5-884c-f35ba1581393-secret-volume\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.337773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e14dd6-904b-45e5-884c-f35ba1581393-secret-volume\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.337825 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e14dd6-904b-45e5-884c-f35ba1581393-config-volume\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.337925 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m44b\" (UniqueName: \"kubernetes.io/projected/41e14dd6-904b-45e5-884c-f35ba1581393-kube-api-access-6m44b\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.339905 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e14dd6-904b-45e5-884c-f35ba1581393-config-volume\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.344568 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e14dd6-904b-45e5-884c-f35ba1581393-secret-volume\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.361943 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m44b\" (UniqueName: \"kubernetes.io/projected/41e14dd6-904b-45e5-884c-f35ba1581393-kube-api-access-6m44b\") pod \"collect-profiles-29423280-kf6qf\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:00 crc kubenswrapper[4894]: I1210 20:00:00.510822 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:01 crc kubenswrapper[4894]: I1210 20:00:01.015508 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf"] Dec 10 20:00:01 crc kubenswrapper[4894]: I1210 20:00:01.262587 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" event={"ID":"41e14dd6-904b-45e5-884c-f35ba1581393","Type":"ContainerStarted","Data":"60d583ee011841f0e64d7e5c2bb71ed638bfe0b09872698f20ba4721023fca2c"} Dec 10 20:00:01 crc kubenswrapper[4894]: I1210 20:00:01.262909 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" event={"ID":"41e14dd6-904b-45e5-884c-f35ba1581393","Type":"ContainerStarted","Data":"a6a63eb35a7a43aa93381b80f4bd546f7a6c2f6a1ae86936eaba95a137b3c186"} Dec 10 20:00:01 crc kubenswrapper[4894]: I1210 20:00:01.288171 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" podStartSLOduration=1.288149598 podStartE2EDuration="1.288149598s" podCreationTimestamp="2025-12-10 20:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 20:00:01.280870192 +0000 UTC m=+3884.075717994" watchObservedRunningTime="2025-12-10 20:00:01.288149598 +0000 UTC m=+3884.082997400" Dec 10 20:00:02 crc kubenswrapper[4894]: I1210 20:00:02.302165 4894 generic.go:334] "Generic (PLEG): container finished" podID="41e14dd6-904b-45e5-884c-f35ba1581393" containerID="60d583ee011841f0e64d7e5c2bb71ed638bfe0b09872698f20ba4721023fca2c" exitCode=0 Dec 10 20:00:02 crc kubenswrapper[4894]: I1210 20:00:02.302215 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" event={"ID":"41e14dd6-904b-45e5-884c-f35ba1581393","Type":"ContainerDied","Data":"60d583ee011841f0e64d7e5c2bb71ed638bfe0b09872698f20ba4721023fca2c"} Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.762804 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.812312 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e14dd6-904b-45e5-884c-f35ba1581393-config-volume\") pod \"41e14dd6-904b-45e5-884c-f35ba1581393\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.812550 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e14dd6-904b-45e5-884c-f35ba1581393-secret-volume\") pod \"41e14dd6-904b-45e5-884c-f35ba1581393\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.812728 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m44b\" (UniqueName: \"kubernetes.io/projected/41e14dd6-904b-45e5-884c-f35ba1581393-kube-api-access-6m44b\") pod \"41e14dd6-904b-45e5-884c-f35ba1581393\" (UID: \"41e14dd6-904b-45e5-884c-f35ba1581393\") " Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.813207 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41e14dd6-904b-45e5-884c-f35ba1581393-config-volume" (OuterVolumeSpecName: "config-volume") pod "41e14dd6-904b-45e5-884c-f35ba1581393" (UID: "41e14dd6-904b-45e5-884c-f35ba1581393"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.813372 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41e14dd6-904b-45e5-884c-f35ba1581393-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.823746 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e14dd6-904b-45e5-884c-f35ba1581393-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "41e14dd6-904b-45e5-884c-f35ba1581393" (UID: "41e14dd6-904b-45e5-884c-f35ba1581393"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.823811 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41e14dd6-904b-45e5-884c-f35ba1581393-kube-api-access-6m44b" (OuterVolumeSpecName: "kube-api-access-6m44b") pod "41e14dd6-904b-45e5-884c-f35ba1581393" (UID: "41e14dd6-904b-45e5-884c-f35ba1581393"). InnerVolumeSpecName "kube-api-access-6m44b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.915155 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41e14dd6-904b-45e5-884c-f35ba1581393-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:00:03 crc kubenswrapper[4894]: I1210 20:00:03.915383 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m44b\" (UniqueName: \"kubernetes.io/projected/41e14dd6-904b-45e5-884c-f35ba1581393-kube-api-access-6m44b\") on node \"crc\" DevicePath \"\"" Dec 10 20:00:04 crc kubenswrapper[4894]: I1210 20:00:04.327251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" event={"ID":"41e14dd6-904b-45e5-884c-f35ba1581393","Type":"ContainerDied","Data":"a6a63eb35a7a43aa93381b80f4bd546f7a6c2f6a1ae86936eaba95a137b3c186"} Dec 10 20:00:04 crc kubenswrapper[4894]: I1210 20:00:04.327353 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-kf6qf" Dec 10 20:00:04 crc kubenswrapper[4894]: I1210 20:00:04.327394 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6a63eb35a7a43aa93381b80f4bd546f7a6c2f6a1ae86936eaba95a137b3c186" Dec 10 20:00:04 crc kubenswrapper[4894]: I1210 20:00:04.370610 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn"] Dec 10 20:00:04 crc kubenswrapper[4894]: I1210 20:00:04.382406 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-zm2dn"] Dec 10 20:00:05 crc kubenswrapper[4894]: I1210 20:00:05.558487 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b191d21-d1ab-4df8-9e29-6747e121a047" path="/var/lib/kubelet/pods/7b191d21-d1ab-4df8-9e29-6747e121a047/volumes" Dec 10 20:00:12 crc kubenswrapper[4894]: I1210 20:00:12.540975 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:00:12 crc kubenswrapper[4894]: E1210 20:00:12.542001 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:00:25 crc kubenswrapper[4894]: I1210 20:00:25.360418 4894 scope.go:117] "RemoveContainer" containerID="f5e8428cc2584c191634b4c72457c154f95fc3661e66f4c4c45a958eb0f22ff4" Dec 10 20:00:25 crc kubenswrapper[4894]: I1210 20:00:25.541655 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:00:25 crc kubenswrapper[4894]: E1210 20:00:25.542219 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:00:36 crc kubenswrapper[4894]: I1210 20:00:36.540659 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:00:36 crc kubenswrapper[4894]: E1210 20:00:36.541631 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:00:50 crc kubenswrapper[4894]: I1210 20:00:50.541623 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:00:50 crc kubenswrapper[4894]: E1210 20:00:50.542805 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.561943 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z8cn2"] Dec 10 20:00:51 crc kubenswrapper[4894]: E1210 20:00:51.562888 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41e14dd6-904b-45e5-884c-f35ba1581393" containerName="collect-profiles" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.562903 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41e14dd6-904b-45e5-884c-f35ba1581393" containerName="collect-profiles" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.563154 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41e14dd6-904b-45e5-884c-f35ba1581393" containerName="collect-profiles" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.566694 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.581547 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z8cn2"] Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.623166 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-catalog-content\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.623260 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8tzp\" (UniqueName: \"kubernetes.io/projected/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-kube-api-access-m8tzp\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.623320 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-utilities\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.725878 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-catalog-content\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.725948 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8tzp\" (UniqueName: \"kubernetes.io/projected/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-kube-api-access-m8tzp\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.726003 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-utilities\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.726345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-catalog-content\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.726387 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-utilities\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.746450 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8tzp\" (UniqueName: \"kubernetes.io/projected/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-kube-api-access-m8tzp\") pod \"redhat-marketplace-z8cn2\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:51 crc kubenswrapper[4894]: I1210 20:00:51.898191 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:00:52 crc kubenswrapper[4894]: I1210 20:00:52.403918 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z8cn2"] Dec 10 20:00:52 crc kubenswrapper[4894]: I1210 20:00:52.811182 4894 generic.go:334] "Generic (PLEG): container finished" podID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerID="5aa3bf1a56f040fe3ab7eea351bdee5e596b41e7659de06716998e532e5b3ac8" exitCode=0 Dec 10 20:00:52 crc kubenswrapper[4894]: I1210 20:00:52.811238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z8cn2" event={"ID":"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36","Type":"ContainerDied","Data":"5aa3bf1a56f040fe3ab7eea351bdee5e596b41e7659de06716998e532e5b3ac8"} Dec 10 20:00:52 crc kubenswrapper[4894]: I1210 20:00:52.811529 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z8cn2" event={"ID":"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36","Type":"ContainerStarted","Data":"6c167f1784235babdcc8724bcad26cf2ae12564bfb1b1b1a30998d63958da81e"} Dec 10 20:00:52 crc kubenswrapper[4894]: I1210 20:00:52.813676 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:00:54 crc kubenswrapper[4894]: I1210 20:00:54.831963 4894 generic.go:334] "Generic (PLEG): container finished" podID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerID="9475c9288de00492cb77200d398811f3263e7c22f645c82c80b4867056f1689f" exitCode=0 Dec 10 20:00:54 crc kubenswrapper[4894]: I1210 20:00:54.832141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z8cn2" event={"ID":"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36","Type":"ContainerDied","Data":"9475c9288de00492cb77200d398811f3263e7c22f645c82c80b4867056f1689f"} Dec 10 20:00:55 crc kubenswrapper[4894]: I1210 20:00:55.845292 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z8cn2" event={"ID":"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36","Type":"ContainerStarted","Data":"efeb530835aa0a262d4eac21c2ed03b7a09c5f379e067f544e23eb5e79cbb11a"} Dec 10 20:00:55 crc kubenswrapper[4894]: I1210 20:00:55.875423 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z8cn2" podStartSLOduration=2.391370123 podStartE2EDuration="4.875405359s" podCreationTimestamp="2025-12-10 20:00:51 +0000 UTC" firstStartedPulling="2025-12-10 20:00:52.813417786 +0000 UTC m=+3935.608265588" lastFinishedPulling="2025-12-10 20:00:55.297453032 +0000 UTC m=+3938.092300824" observedRunningTime="2025-12-10 20:00:55.863726984 +0000 UTC m=+3938.658574796" watchObservedRunningTime="2025-12-10 20:00:55.875405359 +0000 UTC m=+3938.670253151" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.165043 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29423281-wfvxd"] Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.167308 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.187005 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423281-wfvxd"] Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.222545 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-fernet-keys\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.222988 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-config-data\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.223243 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-combined-ca-bundle\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.223477 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6v6c\" (UniqueName: \"kubernetes.io/projected/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-kube-api-access-d6v6c\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.325477 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6v6c\" (UniqueName: \"kubernetes.io/projected/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-kube-api-access-d6v6c\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.325611 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-fernet-keys\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.325638 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-config-data\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.325713 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-combined-ca-bundle\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.331432 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-fernet-keys\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.332121 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-combined-ca-bundle\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.332814 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-config-data\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.343455 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6v6c\" (UniqueName: \"kubernetes.io/projected/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-kube-api-access-d6v6c\") pod \"keystone-cron-29423281-wfvxd\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.518636 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:00 crc kubenswrapper[4894]: I1210 20:01:00.990450 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423281-wfvxd"] Dec 10 20:01:01 crc kubenswrapper[4894]: I1210 20:01:01.898782 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:01:01 crc kubenswrapper[4894]: I1210 20:01:01.899221 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:01:01 crc kubenswrapper[4894]: I1210 20:01:01.904278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-wfvxd" event={"ID":"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e","Type":"ContainerStarted","Data":"3b73a8b495b421fbefe11db41d365b73acd1b647515aca0478d0dc4faacc37c4"} Dec 10 20:01:01 crc kubenswrapper[4894]: I1210 20:01:01.904333 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-wfvxd" event={"ID":"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e","Type":"ContainerStarted","Data":"11aecda5a2f2db706745571bc47f329f921cb0a5e6a7b53a9011656fb33ffd86"} Dec 10 20:01:01 crc kubenswrapper[4894]: I1210 20:01:01.946758 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29423281-wfvxd" podStartSLOduration=1.946740253 podStartE2EDuration="1.946740253s" podCreationTimestamp="2025-12-10 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 20:01:01.93994696 +0000 UTC m=+3944.734794792" watchObservedRunningTime="2025-12-10 20:01:01.946740253 +0000 UTC m=+3944.741588045" Dec 10 20:01:01 crc kubenswrapper[4894]: I1210 20:01:01.995261 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:01:02 crc kubenswrapper[4894]: I1210 20:01:02.987474 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:01:03 crc kubenswrapper[4894]: I1210 20:01:03.038117 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z8cn2"] Dec 10 20:01:03 crc kubenswrapper[4894]: I1210 20:01:03.928350 4894 generic.go:334] "Generic (PLEG): container finished" podID="9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" containerID="3b73a8b495b421fbefe11db41d365b73acd1b647515aca0478d0dc4faacc37c4" exitCode=0 Dec 10 20:01:03 crc kubenswrapper[4894]: I1210 20:01:03.928839 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-wfvxd" event={"ID":"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e","Type":"ContainerDied","Data":"3b73a8b495b421fbefe11db41d365b73acd1b647515aca0478d0dc4faacc37c4"} Dec 10 20:01:04 crc kubenswrapper[4894]: I1210 20:01:04.542532 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:01:04 crc kubenswrapper[4894]: E1210 20:01:04.544214 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:01:04 crc kubenswrapper[4894]: I1210 20:01:04.938147 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z8cn2" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="registry-server" containerID="cri-o://efeb530835aa0a262d4eac21c2ed03b7a09c5f379e067f544e23eb5e79cbb11a" gracePeriod=2 Dec 10 20:01:06 crc kubenswrapper[4894]: I1210 20:01:06.992667 4894 generic.go:334] "Generic (PLEG): container finished" podID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerID="efeb530835aa0a262d4eac21c2ed03b7a09c5f379e067f544e23eb5e79cbb11a" exitCode=0 Dec 10 20:01:06 crc kubenswrapper[4894]: I1210 20:01:06.992997 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z8cn2" event={"ID":"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36","Type":"ContainerDied","Data":"efeb530835aa0a262d4eac21c2ed03b7a09c5f379e067f544e23eb5e79cbb11a"} Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.494298 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.627421 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6v6c\" (UniqueName: \"kubernetes.io/projected/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-kube-api-access-d6v6c\") pod \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.627896 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-config-data\") pod \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.627975 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-fernet-keys\") pod \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.628115 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-combined-ca-bundle\") pod \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\" (UID: \"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.632956 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" (UID: "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.633584 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-kube-api-access-d6v6c" (OuterVolumeSpecName: "kube-api-access-d6v6c") pod "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" (UID: "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e"). InnerVolumeSpecName "kube-api-access-d6v6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.658072 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" (UID: "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.686444 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-config-data" (OuterVolumeSpecName: "config-data") pod "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" (UID: "9bfbab74-f8db-4bb0-9638-4a7bfd252e7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.730133 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.730165 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6v6c\" (UniqueName: \"kubernetes.io/projected/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-kube-api-access-d6v6c\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.730177 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.730188 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9bfbab74-f8db-4bb0-9638-4a7bfd252e7e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.770164 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.831943 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-utilities\") pod \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.832105 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8tzp\" (UniqueName: \"kubernetes.io/projected/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-kube-api-access-m8tzp\") pod \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.832294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-catalog-content\") pod \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\" (UID: \"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36\") " Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.840528 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-utilities" (OuterVolumeSpecName: "utilities") pod "a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" (UID: "a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.840803 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-kube-api-access-m8tzp" (OuterVolumeSpecName: "kube-api-access-m8tzp") pod "a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" (UID: "a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36"). InnerVolumeSpecName "kube-api-access-m8tzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.866006 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" (UID: "a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.935716 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.936062 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:07 crc kubenswrapper[4894]: I1210 20:01:07.936144 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8tzp\" (UniqueName: \"kubernetes.io/projected/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36-kube-api-access-m8tzp\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.002895 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-wfvxd" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.003359 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-wfvxd" event={"ID":"9bfbab74-f8db-4bb0-9638-4a7bfd252e7e","Type":"ContainerDied","Data":"11aecda5a2f2db706745571bc47f329f921cb0a5e6a7b53a9011656fb33ffd86"} Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.003397 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11aecda5a2f2db706745571bc47f329f921cb0a5e6a7b53a9011656fb33ffd86" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.005394 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z8cn2" event={"ID":"a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36","Type":"ContainerDied","Data":"6c167f1784235babdcc8724bcad26cf2ae12564bfb1b1b1a30998d63958da81e"} Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.005431 4894 scope.go:117] "RemoveContainer" containerID="efeb530835aa0a262d4eac21c2ed03b7a09c5f379e067f544e23eb5e79cbb11a" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.005466 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z8cn2" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.024222 4894 scope.go:117] "RemoveContainer" containerID="9475c9288de00492cb77200d398811f3263e7c22f645c82c80b4867056f1689f" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.046076 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z8cn2"] Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.046655 4894 scope.go:117] "RemoveContainer" containerID="5aa3bf1a56f040fe3ab7eea351bdee5e596b41e7659de06716998e532e5b3ac8" Dec 10 20:01:08 crc kubenswrapper[4894]: I1210 20:01:08.058870 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z8cn2"] Dec 10 20:01:09 crc kubenswrapper[4894]: I1210 20:01:09.560452 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" path="/var/lib/kubelet/pods/a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36/volumes" Dec 10 20:01:15 crc kubenswrapper[4894]: I1210 20:01:15.540646 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:01:15 crc kubenswrapper[4894]: E1210 20:01:15.541557 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:01:28 crc kubenswrapper[4894]: I1210 20:01:28.540854 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:01:28 crc kubenswrapper[4894]: E1210 20:01:28.541754 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:01:41 crc kubenswrapper[4894]: I1210 20:01:41.541823 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:01:41 crc kubenswrapper[4894]: E1210 20:01:41.545142 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-vbjj4_openshift-machine-config-operator(3fd55194-d263-4547-9b8f-934ef8050b59)\"" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" podUID="3fd55194-d263-4547-9b8f-934ef8050b59" Dec 10 20:01:52 crc kubenswrapper[4894]: I1210 20:01:52.542475 4894 scope.go:117] "RemoveContainer" containerID="088e72bdf863bacb7e217ff03066c1173054cabd259a6aff0400eaf9852e4935" Dec 10 20:01:53 crc kubenswrapper[4894]: I1210 20:01:53.464828 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-vbjj4" event={"ID":"3fd55194-d263-4547-9b8f-934ef8050b59","Type":"ContainerStarted","Data":"74d4aa33f4ccc5ac6fef6dbd6621eeba02d4297914896e0c5be0523c83abfafe"} Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.471449 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdg5w"] Dec 10 20:02:32 crc kubenswrapper[4894]: E1210 20:02:32.472591 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="extract-utilities" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.472609 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="extract-utilities" Dec 10 20:02:32 crc kubenswrapper[4894]: E1210 20:02:32.472633 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" containerName="keystone-cron" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.472641 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" containerName="keystone-cron" Dec 10 20:02:32 crc kubenswrapper[4894]: E1210 20:02:32.472660 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="extract-content" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.472670 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="extract-content" Dec 10 20:02:32 crc kubenswrapper[4894]: E1210 20:02:32.472711 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="registry-server" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.472719 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="registry-server" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.473009 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bfbab74-f8db-4bb0-9638-4a7bfd252e7e" containerName="keystone-cron" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.473036 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a44fd94c-d6e6-4c7a-a3ca-46ad462c0b36" containerName="registry-server" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.474896 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.482170 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdg5w"] Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.567509 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbx9q\" (UniqueName: \"kubernetes.io/projected/d4310da2-eb80-48b9-9f71-a505d3999ab2-kube-api-access-jbx9q\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.567581 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-catalog-content\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.567839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-utilities\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.669795 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbx9q\" (UniqueName: \"kubernetes.io/projected/d4310da2-eb80-48b9-9f71-a505d3999ab2-kube-api-access-jbx9q\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.669899 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-catalog-content\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.670007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-utilities\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.670424 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-catalog-content\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.670449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-utilities\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.693030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbx9q\" (UniqueName: \"kubernetes.io/projected/d4310da2-eb80-48b9-9f71-a505d3999ab2-kube-api-access-jbx9q\") pod \"redhat-operators-sdg5w\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:32 crc kubenswrapper[4894]: I1210 20:02:32.802947 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:33 crc kubenswrapper[4894]: I1210 20:02:33.336548 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdg5w"] Dec 10 20:02:33 crc kubenswrapper[4894]: I1210 20:02:33.863898 4894 generic.go:334] "Generic (PLEG): container finished" podID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerID="74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319" exitCode=0 Dec 10 20:02:33 crc kubenswrapper[4894]: I1210 20:02:33.863941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerDied","Data":"74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319"} Dec 10 20:02:33 crc kubenswrapper[4894]: I1210 20:02:33.864230 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerStarted","Data":"74ddd7003f116d6e36f8f926cf85a2ed3276395e5acbe8cc71bd95eeb0bc6013"} Dec 10 20:02:34 crc kubenswrapper[4894]: I1210 20:02:34.881764 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerStarted","Data":"e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a"} Dec 10 20:02:38 crc kubenswrapper[4894]: I1210 20:02:38.927558 4894 generic.go:334] "Generic (PLEG): container finished" podID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerID="e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a" exitCode=0 Dec 10 20:02:38 crc kubenswrapper[4894]: I1210 20:02:38.928877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerDied","Data":"e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a"} Dec 10 20:02:39 crc kubenswrapper[4894]: I1210 20:02:39.944227 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerStarted","Data":"4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14"} Dec 10 20:02:39 crc kubenswrapper[4894]: I1210 20:02:39.972028 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdg5w" podStartSLOduration=2.481369033 podStartE2EDuration="7.972010235s" podCreationTimestamp="2025-12-10 20:02:32 +0000 UTC" firstStartedPulling="2025-12-10 20:02:33.86650038 +0000 UTC m=+4036.661348172" lastFinishedPulling="2025-12-10 20:02:39.357141582 +0000 UTC m=+4042.151989374" observedRunningTime="2025-12-10 20:02:39.969014724 +0000 UTC m=+4042.763862566" watchObservedRunningTime="2025-12-10 20:02:39.972010235 +0000 UTC m=+4042.766858037" Dec 10 20:02:42 crc kubenswrapper[4894]: I1210 20:02:42.803559 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:42 crc kubenswrapper[4894]: I1210 20:02:42.804230 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:43 crc kubenswrapper[4894]: I1210 20:02:43.860836 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sdg5w" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="registry-server" probeResult="failure" output=< Dec 10 20:02:43 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 10 20:02:43 crc kubenswrapper[4894]: > Dec 10 20:02:52 crc kubenswrapper[4894]: I1210 20:02:52.888231 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:52 crc kubenswrapper[4894]: I1210 20:02:52.965891 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:53 crc kubenswrapper[4894]: I1210 20:02:53.136621 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdg5w"] Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.089745 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sdg5w" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="registry-server" containerID="cri-o://4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14" gracePeriod=2 Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.715048 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.842568 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbx9q\" (UniqueName: \"kubernetes.io/projected/d4310da2-eb80-48b9-9f71-a505d3999ab2-kube-api-access-jbx9q\") pod \"d4310da2-eb80-48b9-9f71-a505d3999ab2\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.842645 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-utilities\") pod \"d4310da2-eb80-48b9-9f71-a505d3999ab2\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.842699 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-catalog-content\") pod \"d4310da2-eb80-48b9-9f71-a505d3999ab2\" (UID: \"d4310da2-eb80-48b9-9f71-a505d3999ab2\") " Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.849584 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-utilities" (OuterVolumeSpecName: "utilities") pod "d4310da2-eb80-48b9-9f71-a505d3999ab2" (UID: "d4310da2-eb80-48b9-9f71-a505d3999ab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.853991 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4310da2-eb80-48b9-9f71-a505d3999ab2-kube-api-access-jbx9q" (OuterVolumeSpecName: "kube-api-access-jbx9q") pod "d4310da2-eb80-48b9-9f71-a505d3999ab2" (UID: "d4310da2-eb80-48b9-9f71-a505d3999ab2"). InnerVolumeSpecName "kube-api-access-jbx9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.946017 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbx9q\" (UniqueName: \"kubernetes.io/projected/d4310da2-eb80-48b9-9f71-a505d3999ab2-kube-api-access-jbx9q\") on node \"crc\" DevicePath \"\"" Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.946077 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:02:54 crc kubenswrapper[4894]: I1210 20:02:54.979506 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4310da2-eb80-48b9-9f71-a505d3999ab2" (UID: "d4310da2-eb80-48b9-9f71-a505d3999ab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.048427 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4310da2-eb80-48b9-9f71-a505d3999ab2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.102148 4894 generic.go:334] "Generic (PLEG): container finished" podID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerID="4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14" exitCode=0 Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.102243 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdg5w" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.102249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerDied","Data":"4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14"} Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.102806 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdg5w" event={"ID":"d4310da2-eb80-48b9-9f71-a505d3999ab2","Type":"ContainerDied","Data":"74ddd7003f116d6e36f8f926cf85a2ed3276395e5acbe8cc71bd95eeb0bc6013"} Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.102836 4894 scope.go:117] "RemoveContainer" containerID="4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.134442 4894 scope.go:117] "RemoveContainer" containerID="e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.145949 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdg5w"] Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.156081 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sdg5w"] Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.553015 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" path="/var/lib/kubelet/pods/d4310da2-eb80-48b9-9f71-a505d3999ab2/volumes" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.708332 4894 scope.go:117] "RemoveContainer" containerID="74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.774908 4894 scope.go:117] "RemoveContainer" containerID="4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14" Dec 10 20:02:55 crc kubenswrapper[4894]: E1210 20:02:55.776248 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14\": container with ID starting with 4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14 not found: ID does not exist" containerID="4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.776319 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14"} err="failed to get container status \"4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14\": rpc error: code = NotFound desc = could not find container \"4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14\": container with ID starting with 4dea8cdd9cc83f5b754bb809cf61d9ea1a5320315b2be6d480f938f78ce55a14 not found: ID does not exist" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.776356 4894 scope.go:117] "RemoveContainer" containerID="e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a" Dec 10 20:02:55 crc kubenswrapper[4894]: E1210 20:02:55.776811 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a\": container with ID starting with e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a not found: ID does not exist" containerID="e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.776873 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a"} err="failed to get container status \"e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a\": rpc error: code = NotFound desc = could not find container \"e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a\": container with ID starting with e28147132ad5dd546dcf4a8c46be744eeb4332a84e29abf6454a7f4393fb376a not found: ID does not exist" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.776896 4894 scope.go:117] "RemoveContainer" containerID="74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319" Dec 10 20:02:55 crc kubenswrapper[4894]: E1210 20:02:55.777165 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319\": container with ID starting with 74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319 not found: ID does not exist" containerID="74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319" Dec 10 20:02:55 crc kubenswrapper[4894]: I1210 20:02:55.777214 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319"} err="failed to get container status \"74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319\": rpc error: code = NotFound desc = could not find container \"74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319\": container with ID starting with 74791b8bdba8a27753f6c48f4f52f9a21a3dc890427527809ab428b8e8add319 not found: ID does not exist" Dec 10 20:03:21 crc kubenswrapper[4894]: I1210 20:03:21.995807 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mpcq6"] Dec 10 20:03:21 crc kubenswrapper[4894]: E1210 20:03:21.997505 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="extract-content" Dec 10 20:03:21 crc kubenswrapper[4894]: I1210 20:03:21.997526 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="extract-content" Dec 10 20:03:21 crc kubenswrapper[4894]: E1210 20:03:21.997555 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="extract-utilities" Dec 10 20:03:21 crc kubenswrapper[4894]: I1210 20:03:21.997566 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="extract-utilities" Dec 10 20:03:21 crc kubenswrapper[4894]: E1210 20:03:21.997581 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="registry-server" Dec 10 20:03:21 crc kubenswrapper[4894]: I1210 20:03:21.997589 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="registry-server" Dec 10 20:03:21 crc kubenswrapper[4894]: I1210 20:03:21.997910 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4310da2-eb80-48b9-9f71-a505d3999ab2" containerName="registry-server" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:21.999865 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.008791 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mpcq6"] Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.134413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-catalog-content\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.134490 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-utilities\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.134518 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqdr2\" (UniqueName: \"kubernetes.io/projected/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-kube-api-access-fqdr2\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.236400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-utilities\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.236447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqdr2\" (UniqueName: \"kubernetes.io/projected/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-kube-api-access-fqdr2\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.236595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-catalog-content\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.237097 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-catalog-content\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.237236 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-utilities\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.258899 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqdr2\" (UniqueName: \"kubernetes.io/projected/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-kube-api-access-fqdr2\") pod \"community-operators-mpcq6\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.318816 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:22 crc kubenswrapper[4894]: I1210 20:03:22.903288 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mpcq6"] Dec 10 20:03:23 crc kubenswrapper[4894]: I1210 20:03:23.411955 4894 generic.go:334] "Generic (PLEG): container finished" podID="07a484c8-d63b-4e67-86bd-7f5b6bbd2636" containerID="1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9" exitCode=0 Dec 10 20:03:23 crc kubenswrapper[4894]: I1210 20:03:23.412052 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerDied","Data":"1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9"} Dec 10 20:03:23 crc kubenswrapper[4894]: I1210 20:03:23.412251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerStarted","Data":"2d797fc62b638c1a94a17671fd431fbeeb9abd0aaac0cb8af375c7f0aa6bb140"} Dec 10 20:03:25 crc kubenswrapper[4894]: I1210 20:03:25.433704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerStarted","Data":"496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5"} Dec 10 20:03:26 crc kubenswrapper[4894]: I1210 20:03:26.444140 4894 generic.go:334] "Generic (PLEG): container finished" podID="07a484c8-d63b-4e67-86bd-7f5b6bbd2636" containerID="496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5" exitCode=0 Dec 10 20:03:26 crc kubenswrapper[4894]: I1210 20:03:26.444217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerDied","Data":"496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5"} Dec 10 20:03:27 crc kubenswrapper[4894]: I1210 20:03:27.454549 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerStarted","Data":"3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9"} Dec 10 20:03:27 crc kubenswrapper[4894]: I1210 20:03:27.482731 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mpcq6" podStartSLOduration=2.946492633 podStartE2EDuration="6.482704753s" podCreationTimestamp="2025-12-10 20:03:21 +0000 UTC" firstStartedPulling="2025-12-10 20:03:23.414179408 +0000 UTC m=+4086.209027200" lastFinishedPulling="2025-12-10 20:03:26.950391528 +0000 UTC m=+4089.745239320" observedRunningTime="2025-12-10 20:03:27.47370086 +0000 UTC m=+4090.268548652" watchObservedRunningTime="2025-12-10 20:03:27.482704753 +0000 UTC m=+4090.277552565" Dec 10 20:03:32 crc kubenswrapper[4894]: I1210 20:03:32.319840 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:32 crc kubenswrapper[4894]: I1210 20:03:32.320286 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:32 crc kubenswrapper[4894]: I1210 20:03:32.388558 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:32 crc kubenswrapper[4894]: I1210 20:03:32.575493 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:32 crc kubenswrapper[4894]: I1210 20:03:32.633449 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mpcq6"] Dec 10 20:03:34 crc kubenswrapper[4894]: I1210 20:03:34.545077 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mpcq6" podUID="07a484c8-d63b-4e67-86bd-7f5b6bbd2636" containerName="registry-server" containerID="cri-o://3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9" gracePeriod=2 Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.140705 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.227010 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-utilities\") pod \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.227172 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-catalog-content\") pod \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.227220 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqdr2\" (UniqueName: \"kubernetes.io/projected/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-kube-api-access-fqdr2\") pod \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\" (UID: \"07a484c8-d63b-4e67-86bd-7f5b6bbd2636\") " Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.229067 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-utilities" (OuterVolumeSpecName: "utilities") pod "07a484c8-d63b-4e67-86bd-7f5b6bbd2636" (UID: "07a484c8-d63b-4e67-86bd-7f5b6bbd2636"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.233825 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-kube-api-access-fqdr2" (OuterVolumeSpecName: "kube-api-access-fqdr2") pod "07a484c8-d63b-4e67-86bd-7f5b6bbd2636" (UID: "07a484c8-d63b-4e67-86bd-7f5b6bbd2636"). InnerVolumeSpecName "kube-api-access-fqdr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.287200 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07a484c8-d63b-4e67-86bd-7f5b6bbd2636" (UID: "07a484c8-d63b-4e67-86bd-7f5b6bbd2636"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.329974 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.330011 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.330022 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqdr2\" (UniqueName: \"kubernetes.io/projected/07a484c8-d63b-4e67-86bd-7f5b6bbd2636-kube-api-access-fqdr2\") on node \"crc\" DevicePath \"\"" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.557352 4894 generic.go:334] "Generic (PLEG): container finished" podID="07a484c8-d63b-4e67-86bd-7f5b6bbd2636" containerID="3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9" exitCode=0 Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.557526 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpcq6" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.574107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerDied","Data":"3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9"} Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.574155 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpcq6" event={"ID":"07a484c8-d63b-4e67-86bd-7f5b6bbd2636","Type":"ContainerDied","Data":"2d797fc62b638c1a94a17671fd431fbeeb9abd0aaac0cb8af375c7f0aa6bb140"} Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.574219 4894 scope.go:117] "RemoveContainer" containerID="3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.605058 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mpcq6"] Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.607839 4894 scope.go:117] "RemoveContainer" containerID="496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.615137 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mpcq6"] Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.634541 4894 scope.go:117] "RemoveContainer" containerID="1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.681289 4894 scope.go:117] "RemoveContainer" containerID="3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9" Dec 10 20:03:35 crc kubenswrapper[4894]: E1210 20:03:35.681647 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9\": container with ID starting with 3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9 not found: ID does not exist" containerID="3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.681674 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9"} err="failed to get container status \"3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9\": rpc error: code = NotFound desc = could not find container \"3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9\": container with ID starting with 3879d5c3d2ed165ddf022ac2fd51c6c6df905a59858455e18406ff5118f518d9 not found: ID does not exist" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.681694 4894 scope.go:117] "RemoveContainer" containerID="496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5" Dec 10 20:03:35 crc kubenswrapper[4894]: E1210 20:03:35.681960 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5\": container with ID starting with 496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5 not found: ID does not exist" containerID="496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.681982 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5"} err="failed to get container status \"496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5\": rpc error: code = NotFound desc = could not find container \"496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5\": container with ID starting with 496165866d7968d42849c22fc6339af6b468e901a81d95b4a6aff17a39ca47c5 not found: ID does not exist" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.681994 4894 scope.go:117] "RemoveContainer" containerID="1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9" Dec 10 20:03:35 crc kubenswrapper[4894]: E1210 20:03:35.682241 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9\": container with ID starting with 1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9 not found: ID does not exist" containerID="1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9" Dec 10 20:03:35 crc kubenswrapper[4894]: I1210 20:03:35.682284 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9"} err="failed to get container status \"1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9\": rpc error: code = NotFound desc = could not find container \"1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9\": container with ID starting with 1a987cf347c25ec826cca6e64bd0ac3f33f168c18409ea1eeee7a293b7301ce9 not found: ID does not exist" Dec 10 20:03:37 crc kubenswrapper[4894]: I1210 20:03:37.554399 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07a484c8-d63b-4e67-86bd-7f5b6bbd2636" path="/var/lib/kubelet/pods/07a484c8-d63b-4e67-86bd-7f5b6bbd2636/volumes"